var/home/core/zuul-output/0000755000175000017500000000000015146012417014526 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015146023271015472 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000320743115146023206020257 0ustar corecore&ikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs$r.k9Gf8^&\u|l&U狿li//|y-o\׼ٮ֜V˿B oo-q3f2f(_}^ſq^?-}ԏ|W E^ <]]ϗj;Hw70g"Gǯ/7ݮfGR)$DD D~m\rٲ] g $\xeK49)ol7(ӗofՈ_k0mZpPneQn͂YEeeɹ ^ʙ|ʕ0MۂAraZwR}@E1.aLk>0M.*x6Q#%q^Hjf4}&|dd#)3c 0'Iw A57&Q"ԉQIF$%* tJ,w˼љ:72 2Zi}e0ʡPBU[.z{&:ɴ@=yN,:;zuL{+>2^G) *ɚL}ӄ]C }I4Vv@%٘e#dc0Fn 촂iHSr`uXWnyMfNQ8M%3KpNGIrND}2SRCK.(^$0^@hH9%!tasKZiu}THW{y|gg WKPW*g,Z0>?<{r.:;.뙘 A|==-$JRPœ*fOԼf^`ig7!)&c(z$5jlUi_η*t:%?vEmO5wtqÜ3Byu '~qlF?}| nLFR6f8yWWxgg ;k44|Ck4U@`i7wFP|؝5g.Og \r1hɨ|аK"B[T'A25T jzdz6"ٍ߬i3x-QI)k=-\$)'/N̔ڧO1_9C'/-#\1, Gꨦ#Jҡ/KqG!(8l KVKx*lJ3os`՛=JPBUGߩnX#;4 O2{Fݫr~AreFj?wQC9yO|$Uv^Z5gIfz}|]|[>ӸUKҳt17ä$ ֈm maUNvة|]k wW-+b8_b|r wFuRI%TѬza\_3/wn>- MR5Ra 5ܫa!뭠vKtr?tRWU1o6jjr<~Tq> `=tJ!aݡ=h6YݭȾju\0Ac/T%;m]~S`#u.Џ1qNp&gK60nqtƅ": C@!P q]G0¿Øp*xYPLSMY 9J}t/A`*t) OW}\+`peGVق?>6E4)%LJno#ˏl_}?1zngbߧ\I;t.U&hoP~(*ט*yw&7ԟ-\صp&NI%`t3ViӽLb㸵2*3d*mQ%"h+ "f "D(~~moH|E3*46$A h}e¥v`tYϞnTq&zcPj(PJ'ήYQDF;?>6z=5t;.J|$x0w#S )$ {SG!pN7,/M(.ΰdƛޜP16$ c:!%Piocej_H!CE4Cb#O\9fcۼ=z)d]g =OS2[3(/hoj$=Zm Mlh>P]}p TRi*KsmM+1 믡`2\fC$Mj-Kp`zbbq$Igǽgr&P29LcIIGAɐ`P-\:BPS`xiP(/T)#iax[|mi FW ʹo7T7뇴热b!g K#XoV甬6xڂ I &m>AtĘud7"3F´[5mǭ*( :{|B_3YK]7-4mοnk'g'{97_e5{52{s4I#n<薗,*e'&Ж0(ݕ`{azYsu/x)W>OK(BSsǽҰ%>kh5nIYk'LVc(a<1mCޢml_t|8Qz:^SĖ[Sm5+_?&cj.i ˿7^1+]h,*aklVIkS7d'q 甗yLkS :}%J6TIsbFʶ褢sFUC)(k-C"TQ[7j39_ɣ4cY\[|Xs܁dIvT-n,'}6ȴ .#Sq9}5:ްPIGc/ҭ٫3@/%hJKZ|Q;|m@}Ym*H@:FUф(vcD%F"i ' VVdmp%;R8P󦟶U"c&,sw~EQ(S +mpN, Mq 70eP1ZK(O5dc}QQufCdX($0j(HX_$G0gLNO mR!c¨h=z,ffK2Zu8C1}PcIr.e'.I(NJKڨzHJodo\4"+(Nck!a}P `x ăil_=!X;__ZNۈO;j -ݙrbH9t֥s`Mh,욕v`;VI&crgB= FYtЩOte=?>V&{Hkj.R2>e3VWX@&avF묇cTy^]m -ĖUP-\!3^.Y9[XԦo Έ')Ji̯H4~)([!.;嶑, }t&&\5u17\I@ 5O? ʴ(aPqPϟ'+'h=TԫeVިO? )-1 8/%\hC(:=4< ,RmDAWfRoUJy ŗ-ܲ(4k%הrΒ]rύW -e]hx&gs7,6BxzxօoFMA['҉F=NGD4sTq1HPld=Q,DQ IJipqc2*;/!~x]y7D7@u邗`unn_ư-a9t_/.9tTo]r8-X{TMYtt =0AMUk}G9^UA,;Tt,"Dxl DfA\w; &`Ͱ٢x'H/jh7hM=~ ֟y[dI~fHIqC۶1Ik\)3 5Ķ']?SؠC"j_6Ÿ9؎]TTjm\D^x6ANbC ]tVUKe$,\ܺI `Qز@UӬ@B {~6caR!=A>\+܁<lW Gϸ}^w'̅dk  C 7fbU{3Se[s %'!?xL 2ڲ]>i+m^CM&WTj7ȗE!NC6P}H`k(FUM gul)b ;2n6'k}ˍ[`-fYX_pL +1wu(#'3"fxsuҮױdy.0]?ݽb+ uV4}rdM$ѢIA$;~Lvigu+]NC5ÿ nNჶT@~ܥ 7-mU,\rXmQALglNʆ P7k%v>"WCyVtnV K`pC?fE?~fjBwU&'ᚡilRї`m] leu]+?T4v\% ;qF0qV(]pP4W =d#t ru\M{Nj.~27)p|Vn60֭l$4԰vg`i{ 6uwŇctyX{>GXg&[ņzP8_ "J~7+0_t[%XU͍ &dtO:odtRWon%*44JٵK+Woc.F3 %N%FF"HH"\$ۤ_5UWd̡bh塘ZRI&{3TUFp/:4TƳ5[۲yzz+ 4D.Ճ`!TnPFp':.4dMFN=/5ܙz,4kA<:z7y0^} "NqK$2$ Ri ?2,ᙌEK@-V3ʱd:/4Kwm2$'dW<qIE2Ľ)5kJҼMЌ DR3csf6rRSr[I߽ogCc;S5ׂdKZ=M3դ#F;SYƘK`K<<ƛ G׌MU.APf\M*t*vw]xo{:l[n=`smFQµtxx7/W%g!&^=SzDNew(æ*m3D Bo.hI"!A6:uQզ}@j=Mo<}nYUw1Xw:]e/sm lˣaVۤkĨdԖ)RtS2 "E I"{;ōCb{yex&Td >@).p$`XKxnX~E膂Og\IGֻq<-uˮ◶>waPcPw3``m- } vS¢=j=1 W=&;JW(7b ?Q.|K,ϩ3g)D͵Q5PBj(h<[rqTɈjM-y͢FY~p_~O5-֠kDNTͷItI1mk"@$AǏ}%S5<`d+0o,AրcbvJ2O`gA2Ȏp@Z#"U4Xk1G;7#m eji'ĒGIqB//(O &1I;svHd=mJW~ړUCOīpAiB^MP=MQ`=JB!"]b6Ƞi]ItЀ'Vf:yo=K˞r:( n72-˒#K9T\aVܩO "^OF1%e"xm뻱~0GBeFO0ޑ]w(zM6j\v00ׅYɓHڦd%NzT@gID!EL2$%Ӧ{(gL pWkn\SDKIIKWi^9)N?[tLjV}}O͌:&c!JC{J` nKlȉW$)YLE%I:/8)*H|]}\E$V*#(G;3U-;q7KǰfξC?ke`~UK mtIC8^P߼fub8P銗KDi'U6K×5 .]H<$ ^D'!" b1D8,?tT q lKxDȜOY2S3ҁ%mo(YT\3}sѦoY=-- /IDd6Gs =[F۴'c,QAIٰ9JXOz);B= @%AIt0v[Ƿ&FJE͙A~IQ%iShnMІt.޿>q=$ts,cJZڗOx2c6 .1zҪR "^Q[ TF )㢥M-GicQ\BL(hO7zNa>>'(Kgc{>/MoD8q̒vv73'9pM&jV3=ɹvYƛ{3iψI4Kp5 d2oOgd||K>R1Qzi#f>夑3KմԔ萴%|xyr>ķx>{E>Z4Ӥ͋#+hI{hNZt 9`b˝`yB,Ȍ=6Z" 8L O)&On?7\7ix@ D_P"~GijbɠM&HtpR:4Si גt&ngb9%islԃ)Hc`ebw|Ī Zg_0FRYeO:Xy_ XC.l.;oX]}:>3K0R|WD\hnZm֏op};ԫ^(fL}0/E>ƥN7OQ.8[ʔh,Rt:p<0-ʁקiߟt[A3)i>3Z i򩸉*ΏlA" &:1;O]-wgϊ)hn&i'v"/ͤqr@8!̴G~7u5/>HB)iYBAXKL =Z@ >lN%hwiiUsIA8Y&=*2 5I bHb3Lh!ޒh7YJt*CyJÄFKKùMt}.l^]El>NK|//f&!B {&g\,}F)L b߀My6Õw7[{Gqzfz3_X !xJ8T<2!)^_ďǂ.\-d)Kl1헐Z1WMʜ5$)M1Lʳsw5ǫR^v|t$VȖA+Lܑ,҂+sM/ѭy)_ÕNvc*@k]ן;trȫpeoxӻo_nfz6ؘҊ?b*bj^Tc?m%3-$h`EbDC;.j0X1dR? ^}Ծե4NI ܓR{Omu/~+^K9>lIxpI"wS S 'MV+Z:H2d,P4J8 L72?og1>b$]ObsKx̊y`bE&>XYs䀚EƂ@K?n>lhTm' nܡvO+0fqf٠r,$/Zt-1-dė}2Or@3?]^ʧM <mBɃkQ }^an.Fg86}I h5&XӘ8,>b _ z>9!Z>gUŞ}xTL̵ F8ՅX/!gqwߑZȖF 3U>gCCY Hsc`% s8,A_R$קQM17h\EL#w@>omJ/ŵ_iݼGw eIJipFrO{uqy/]c 2ėi_e}L~5&lҬt񗽐0/λL[H* JzeMlTr &|R 2ӗh$cdk?vy̦7]Ạ8ph?z]W_MqKJ> QA^"nYG0_8`N 7{Puٽ/}3ymGqF8RŔ.MMWrO»HzC7ݴLLƓxxi2mW4*@`tF)Ċ+@@twml"Ms>\΋"?|NKfֱn !s׭dֲcUh=Ɩ9b&2} -/f;M.~dhÓ5¨LIa6PnzɗBQiG'CXt!*<0U-(qc;}*CiKe@p&Em&x!i6ٱ˭K& FCfJ9%ٕQ·BD-]R1#]TROr}S [;Zcq6xMY 6seAU9c>Xf~TTX)QӅtӚe~=WtX-sJb?U'3X7J4l+Cj%LPFxŰAVG Y%.9Vnd8? ǫjU3k%E)OD:"Ϳ%E)=}l/'O"Q_4ILAٍKK7'lWQVm0c:%UEhZ].1lcazn2ͦ_DQP/2 re%_bR~r9_7*vrv |S.Z!rV%¢EN$i^B^rX؆ z1ǡXtiK`uk&LO./!Z&p:ˏ!_B{{s1>"=b'K=}|+: :8au"N@#=Ugzy]sTv||Aec Xi.gL'—Ʃb4AUqػ< &}BIrwZ\"t%>6ES5oaPqobb,v 2w s1,jX4W->L!NUy*Gݓ KmmlTbc[O`uxOp  |T!|ik3cL_ AvG i\fs$<;uI\XAV{ˍlJsŅjЙNhwfG8>Vڇg18 O3E*dt:|X`Z)|z&V*"9U_R=Wd<)tc(߯)Y]g5>.1C( .K3g&_P9&`|8|Ldl?6o AMҪ1EzyNAtRuxyn\]q_ߍ&zk.)Eu{_rjuWݚ;*6mMq!R{QWR=oVbmyanUn.Uqsy.?W8 r[zW*8nؿ[;vmcoW]"U;gm>?Z֒Z6`!2XY]-Zcp˿˘ɲ}MV<в~!?YXV+lx)RRfb-I7p)3XɯEr^,bfbKJ'@hX><[@ ,&,]$*բk-Yv5 '1T9!(*t 0'b@񲱥-kc6VnR0h& 0Z|ђ8 CGV[4xIIWN?Yt>lf@ Vi`D~ڇŁQLLkY <ZPKoma_u` !>Z;3F\dEB n+0Z ?&s{ 6(E|<ޭLk1Yn(F!%sx]>CTl9"و5 |ݹր|/#.w0ޒx"khD?O`-9C| &8֨O8VH5uH)28 Ǿ-R9~ +#e;U6]aD6Xzqd5y n';)VKL]O@b OIAG Lmc 2;\d˽$Mu>WmCEQuabAJ;`uy-u.M>9VsWٔo RS`S#m8k;(WAXq 8@+S@+' 8U˜z+ZU;=eTtX->9U-q .AV/|\ǔ%&$]1YINJ2]:a0OWvI.O6xMY0/M$ *s5x{gsəL3{$)ՆbG(}1wt!wVf;I&Xi43غgR 6 ݩJ$)}Ta@ nS*X#r#v6*;WJ-_@q.+?DK១btMp1 1Gȩ f,M`,Lr6E} m"8_SK$_#O;V 7=xLOu-ȹ2NKLjp*: 'SasyrFrcC0 ѱ LKV:U} -:U8t[=EAV$=i[mhm"roe5jqf$i>;V0eOޞ4ccc2J1TN.7q;"sդSP) 0v3-)-ٕAg"pZ: "ka+n!e߮lɹL V3Os\ဝ+A= 2䣔AzG\ ` \vc"Kj61O Px"3Pc /' PW*3GX liWv-6W&)cX |]O;C%8@*Z1%8Gk@5^NtY"Fbi8D'+_1&1 7U^k6v읨gQ`LRx+I&s5Www` q:cdʰ H`X;"}B=-/M~C>''1R[sdJm RD3Q{)bJatdq>*Ct/GǍ-`2:u)"\**dPdvc& HwMlF@a5`+F>ΰ-q>0*s%Q)L>$ćYV\dsEGز/:ٕycZtO 2ze31cDB/eWy!A/V4cbpWaPBIpqS<(lȣ'3K?e Z?ڠ8VSZM}pnqL f2D?mzq*a[~;DY〩b𻾋-]f8dBմVs6傊zF"daeY(R+q%sor|.v\sfa:TX%;3Xl= \k>kqBbB;t@/Cԍ)Ga[ r=nl-w/38ѮI*/=2!j\FW+[3=`BZWX Zd>t*Uǖ\*Fu6Y3[yBPj|LcwaIuR;uݷ㺾|47ߍeys=.EinE% 1zY\+͕߬VͭW_겼cazyU1wOw)Ǽn@6 |lk'Z|VZpsqL5 څB}>u)^v~,󿴝} 3+m𢛲Pz_Sp2auQAP*tLnIXA6L7 8UgKdT)*7>p{Pgi-b)>U6IXabPde Ӽ8Ģ8GɄnb'G ֤Mcv4?>HC78NE@UMc8>`TvZ:}O K?m#+,}y_ p,;j-[/8UJ'\AI)R4aL!py3!oO0!vP% d|SRmSRvbtSRv +#3eFF V^zh1i,F+ݗŕcv$]䍀zb#xY\zd~46BK8j.q#aG<Ф54Yu{Yъgj,{-ḃݕ zߖ/Gg\؞]!VA(la{"xa`w'Yd2xAǮnY^hp?#iZF$X%'vK'8+) ƄKׇ4t!tG[|(t 458, #yʋt{uൺޤB s3_~ϦSS7glڟu {?DhfNò>hAa+(oD1`כ#6rr}z@rm=V_}zqJ7K7]2H/r1Pj!?@%[]x)2}B-%<.os,ZQ=s"3,Pt0)yߔ@?'48P@faz]z1NTbZxmBZ2.kx=]=j(Q,g3#]yfYF`~$e3G8E (4|:hg$x͐GAƁ'Eyu#U&[SIdY 2K^`<2O _l= 󳼑C?lO O4m͚VƩd&<E^)0k)^MIKbj7 ONP˘Oꅜ֤*d^_G0OQoXK)'.׳{+ռfSg[ul$g13lSAZinHM4eouuYRL[U/N~X4_@ $j/K xh㝥OS:+c{@!8! b ~A7,'Ys^wyexsb37z|F@ |ٞ{5?r0Msr 7NQĝ G2hDJ20ig 3g˥q\ z c֏x v 'ٛ$o 7w7'dz$ 72^?~  "%{f/&2\ {3t;KHSe ~HxvAAت{oASdٿ>#!r',ug>ٶ|x [Wavgn5t_O$wwu()hPw`E]yKyIYY!Ak~ ; 0 CiP45v;$ r9]4%|p!a8 Uȋ~u{[mKh !xheS x.~|i@}F;$u[i{/UrJf$Mź'1:eSt 6}yzy4sG*޶{ۘ3IC"_|U\1SLݳ ĻpC)[7,((+{7;=. *K93],.6inĺ=B3?yS\fyMVYUBno'R_~y nM"Ϧ$gc}SwƦ,=,.yy} O8aFd6+$N~z뉦q!оG$ju%I~-d$$ 'o&m7W0O=H(޻.xqEK4 ]'"S zu!v-E_E<"0h<|dzECbj$`cG.Iro45>%ъEa4}<3yŢ=}|4+^FDMp흈h<}RE+<Ų L! 4F7TFh+&rvUEn"5CW94ѽb;qv5[OXXSAguhTfoL\,*/z7|n?noy)r)i)c4\ӌᚻR#rYXfOghG @K|?@LSTV\yxÌHUSF<=ai¨ŠrF碍Ra*mNt]C"|.U-oҚߝr{e>!p*3<j^u &$5N ا!D ~zHg܉lwt>UgCe[Yl9j)+B*)gRX쑐oH0C{50l@4iH 1=/a_[K"+)kU[`&)V<wrfʲa`*϶ b32SM\k@LՆ7Ķ`9* Aۚjix\Ke03=9☞V9>4aUk^h([nla5aZ@Y FfSTs+V L[M%,гQ<'jl]E @&@U`)(c1D Z7ւ8w5\9QS-ӚIC1GutnS^MגrTxZbEC>*Za >GGVV~.gyXhK{kV n2絅*}ۑ(E·òYifa`YrN\&GUGv!w![k,j7+!k*j}Q ) 7dXpAjh5-%.S3@qP&R,<:Yri cW_Oq9DqU7Sp!EtLJtٮnx|7&:p@׹d"=#.eK#OIt]gGbGO%˶D[V3;s"wh;D*ɃGO.u9TziJ,AR@:>놱L+KeT+izEpd*yZAA1R}0\4I2DSGwo3]o3mg0Ѕtnp3tռ;flmci X( 3ysVn G;q̴1e<Έo2͸]b$}ɌQMl@h&0G2mv=IpI2 81Lmeg`8pcf8Fwwg7.,<"fO=@a~0Gaml6S?Y̲QLYcSmq7*S۲=2_KMd[ aў ~"Cc[ fqtf$lqhf ,ΘF30}A' 쪿7\DD,`D((`Y&}۷ SމG !*OFP4@F]]t+&%!ͱc6!c:Ff3Jsug$#h剷ᚇW4H1wF2YH},8~wɛ;^Ew[AhSrlCз`[5Xɯ(Cr ģ(L7G4}cV<oyn5Wz~`pU8c=4O+$@ٻm&`8u %sB?tٶOǕ|F6ewuYiiP8WyO{9^744n6ű@\6l^Hϵ3LCH~?Kn@5ޑ@hu(m^kw5 soJ#㙜*X<E㦤S-?Z0GQ_/KC[7 Qg[=kD~_mxᣓF]A%p9^ n#F`W~vJ 88/Ae csn 'w.-}(}ra6Pڏ;çS 0 ` o~J㒮PU~81 #NWN "C.`hxw/h8KCPۂ^rpۮ={h?b4!;{1rd 4Ԃ~ܜnUqkI=ze9=jV0z8a>8|dȱ 7,@ҏ 19HG pg0].6|=tln`$3bjo(E!y?o먄D&GKJjLG+iVLG^O~9s,A $sRRJ4u})5[;Pm My)qeFa@hD ouj.u%it82{Pe'RShg͓ӗpJC5<7"3fndr7aȎ\cb7evxD|XցZ@UZy`Yf2#`wѥ/EroiQޣ*2mfTMq|$o9;EUCWgl27_~鼸$`T\|Jb-+CC-8A{K˷ =וW-EM]ctDE^R0x1G<37i:v]¤e܍4`P6,3Ze(yR]LXx[9^ᄁJQU<0huiY+}2]t{ H~E74jyǩƩK5BBu}Y.a-e7g[> 6ӀY@TM +a `r~H7ls>f]uo>E3!jW @Lvf)2od0/rlq: q5F~~$h  4Q"\1Qd!u,MU*Ժ T B aV/v`ҫIt UUS?VȝU0}Y%n j]M7CfkyFaԞ* jx AT>6LmnfNg+ XQw =:Vmeև=^|$Y=̶ӤVNv+Eם8_-x ݺ\'@~NoIHeR.8*NUI]<hPx<Պ6T' &-/??2}JW ;Jw'>P'tB׆2tAOUҝu}`w}c2lZlwZHe{Rv ^%ށP{wBjI|P|wB{w Y%فPgwBI*>PwOBV v ۝Piz{@JO#ߓPBUB v'4x]%9B_Zw9=Pn!uF<8:wyyq=FTp2qUEl~S,$TpdRIwTIz+LxaҵM_c@oK:QH6I3|`yA: ^Ṣ,OwUShN^!2Rs/U%UXbǴKeAz|~>9$G<_<U`nWE<EK.qJ/ gePպxsE?4eVS2 4[n1.Z1j? ,!jWi`2K4D_`UQ؆C Zz$_b{(x," MG_ |zxa-NIUj(o$a 3Ԑ7[*@5mwg(ߍ rDj@|H4mYُ=5V;,2W>P't-&0uWE* eIco5ίͯ^qQ)C틇 "evkb2Y+ij,k tD'v MEL:W0YH3;|]8M(97QO& Aÿpfm@ŘXˁߔQ"Y_$ʦҷ |J} H+ E`rv^P R-C`P߼i@s9Y3B;+-\Oc WZH3wkJHlh:fͣ"9ϒldVǭ`'/EΪ%Y}0<9myݹYEm_"ELFq߲ږbJ%DBײMOW6{i!Y49z*Qe4@Oi-7;8SBwU +p*D"M?-T 32=,XY&"RAc$>f Zyfu }@¨S4.L( ;b6TJ"BF܀X![ :|0gX/̓s{)KƪCW.r &:V=daJ+V‚f-/g`)2r(o8ɼ8v(n̚ +2ọ&GFgI86Xܵ$->yT!lYpU3)mx*۴85Etzt,Ghe-SXyM׭9EE9ƋDAsD*N.NwQދ~3{`4xNdӠ{ؐ,ġAO4J-q*d6xKS)k9Vb?qLYy̺ @u-yicM讋Ic+Fu`c6#E1mv,ٱ?k%2{L}ߌVJ?>03sckFD1E(Pv/(ҙ >[N =DhD]!m˃sP {=SvhJPty8m߉_ tKP?pR' F<@[Y5c%e w-~0)݆wem%IW }10e4g\%Z"iV}")R%vQʼnmvޣ̌%O {Ū|e7?~g&IϿ^n 9> 6nǕr_]}/{+Z?l v>u?{ꐕE0Ț)Km,*-g~ ,3: ƶ04SG{lϰg/vZiո<>Zib)WâT[rKگBm]/͓`8ӎ0HD712U2X&0| cDY'{lWF(Rs5)njCrֳtRQmoז?a,u,)|0 Ѥb2:T/8^2kO~XbQ bpv$Apr<A)aҘuʁ˒=nȽt]OG-f@|xU*{`%o .3]Bd)Fϒ<x1֢,~j.[&8!8)g:"@]p~crSg+m|aZ2!(W3bc̜1<, vn#IMHc%r,D_btL{[tKlY,M؅C*DuLY2:AX\c VS80&d)0jlfW[)}]\"D D'R0c=t^ӥYiNUyzWOO$8 u 0uzV1=Wcu1X d.m̎ Ӳt9S S}37H5+rڿ(~?=>pFϣϘՈ1n["͡hsQE31$8*w͌/z3)K|g!"i)&!2.nWn" \5-[YsksB[Za8,44ax/DQ2`òJ%KRPn\N X43S0Z _*q8b,ٽߑ8أm6X8 p&NeE9GsՕCϺx eT!"W!b eIІ}ʴ&78cùΏ,m;GR Z}Tt祳co+dj2r8 "T\Q}}}2]%U׻2Ap^ %TMEB4y9Lq6S%eH%zw= {%6 jY8BWx!)pJ~g]XT9 t)OvAw$S\"K,Ƽ[\" 뙗 ;1D3~Unzh'@sp7NEW?t16@J`,?P('Q MkpUe4L +t9̪Ęmcl ferK:įa=;3vfjV[5t,6aÅ"3ppIpR Ƅ9vR՞5)G5y({4ǶuրIAeF3/e>!(^F fa\C'X XPjiAQfܟyDGyr ȟLq4$*\x"<,<?>֣[/L8?dgs.G7bQ8d&WMW,ꐒ#3uwIp 71|g~u,[$80ri*uĺfQ*I\cؼ$ a#x [C%JIZ0X:g;m 󱊈^~UڼΌB*?y8z [GZB2VTx%~+g墩Z7X;B`1#KVMuJ KgzS rlIQk U*u|#Mb+Yj8",Y.7$8j;#=f;ɧ!rVR׌n3aڏhniQ[ofULޘz\4kWev 2K K9#G0˺%`%d!pK0:l 'yc^FiZc3y)>K':d6#&BHcPj+<‡\P <#c_CqL~}B׳"39 1N0l܇6ɱj3'qpukX4j.#ˍcb%c* 6CW. OPb>Lg&ORlEVxHL(kXް;cy֪fV 'jX~Rˍ팃йeZy>`8ZRbdl8Ϝ4JbB25̓T _אm]VA*_i|^DVXk[z%ƒptC+CE?jaEÙgU |[_aS`if= K]=d "g#צ .z<$j%k >Xb9rw;u8I̬"&~;?ܚD`}`SPQEfTT#5&֪cAc,zy9 4?-wpd,@ĬĞ< j@,5Z8=/yj쬟UT3aD5grzRǔii ~p5RFIhU|y3a~priʱ#\QG&ywR«f=`[)$^Jp 0μI>Y%yGzLw8낈ŬZ'2z۵K`ELӷ4~bQ )̐C=0ة\:hp:1kpE^vap;(V&FK}Leנ r NM&) 4n}ڲe-VKkר[).w6jKQV෺[`7/#k^K+XmFMKr9S&Ӕbp*sK*łja[[MvfIpo@?nAMiH)9 1x ԈoH ɹ8#R.dbƶj("/1GJoHpܗsi&w:P't1Ì?,\Ȱ-5 )2 ̫B]^XpF;ۖ|V% ~^C'VLl IojrpDirqK Ѻ VG"tEOZ<2udTfOhߘi1%2 ̰<-ӏt.%6?S/ 0rqx[8xݥ'Qg9r~#s"}O1DxK`T1`%ASʹB&~;t#K48j=8g2 H)bACm8JJH(RW ɻ/E"W.#~.^堨Ndh- q2%@E<xdcT^3;,G~GLE+s.dcޅ%I8=f9}Ϭ4A\3WOV̋zlWGH;( @DC}|p \0Sˬvpc2+~ vVM쒽a.eZ9 k/"ѥ,%ziq6$ԧh)Kٕ>]~2 J^on{BW;Q$'}8QI}YIL j(;`嵘U^PhyQ dSkcӘVlm0hUPL4X(tXS}~0ҡ.tndtT !ap #8!RcFcbu'qpR{JQ|JMx޽ZK}2Mdi~ևC~*~$4 qo_?~T7_~%cZ{8_lג 5n8ZEj9"V!)Dtל:U]-1݋-e ,/CL+C}fKPPCAR,>23wuC@L[&Q(hf5Q~v\܋']ry)O<z8D`koKwó^B#q;Yi9In&)Bc #/p-^?Ljm:SqFh1*udk.uﷃ޸g0>a3eXEe)OGehN8"M1 h8Bqy{5($F7.h$9rh u-uཫϽUV^D<8IJgJLIpxFozH3ױ{77h8'iљˇ7E~;WM=981Hz -eW0FI4UߟI~7>}>d8+N70Oɨ:ƈ+z1MwVg!.f;j֤_];E/|3oƦD#a/VL/3Ѱ_/LM?.ͱӵ'ƭެJ1XS0!bGݙ)0oon"(^K%%Eqf*h`d`S;H̍q)Ter-ZbV[qe3:|2~ϞTwRQ\f '5"/`-|(z/K+7/Kz34o+rܛ!pW 46+De|ҟ@:YO;q$eيv7U/NW }U6XNѓ?z@98y/`F+h ! q*Q,:n?Z!7ƅi+I 7r8bT4L`LnR8S.1՟DPKR#I,5zwC2Wo@%U9.:ٷOҫ^]bea1>H'Si J N -sZ)Z.yJpYl|Y#-++]f'kj#zyz4-_j/k^wј߁Fg$to&pC'3kpіN.0$DbrCE4<_Ij_fggytgosKiOpp(Oxl@9Hޫ tp+,ttPܓS̲V0nX>0rg68: -?AHn{4T&+"w8=r2炚(:@ن~qޥebuO\w^ާxƼd!xu\l\gb  g VٛR[*܌Vm%HkC'Cb/+[Gֈ<h͸giym b!9dQ̋\{4y_GDY}/Ƒm{׵g?'_ysa#t;YHrt_y#fo~=#XK\~דv6 ɊuI|ov10Ro5)˧9*s /.?W}%7x"JeQ-Y$ p=lt{9IѽA/_nBe$ }EݟGv2*ҁ~kVK.a_͹ T2X3JSM^S8s$rs”GBO쿋얧gyogOlQ:7{Mh;7PSw&[;f0jg[3-5CzP$,i RjRR5 MS+0\ G88<è8NW~INI}+'|۳QS<9+QW|j߻O9BܮS .W'MfwaJZt_D&9J[uK?r/{3-?9Thx9ٚJGø6M^k\`Z S")᜺ ;79.l(42!bxo! mhi) Wڜ)|0[Iۥ5 H!-=R;b֍`d-zp}+8gWwRHEz+.i;zr`Ns so YaiqܦLZx\n9\ MtpKlr15exlr=%3\O\MK+A!+dr197N$ QGy.*q#ь&X@tY՚"kH= -l8%!EX:U9%;Њ؂R_ TOH VKM {h&x^,Yx.׎`/zB܋FnS-G GKh?B>7{;N LˈTJd6P_ƚQ#ESXCL0% ( mq)k+8NLo9b!`3`\#-ɮs8>`Bk2> 8XC`)1]0)&A…cs9TE\4Ov1jdBe x+7qϠPa#+I{ϼ4 - 7J$hK23bsfn`ś)hbJL1hN:'E99&#fe?{O۶_!á3.^Ii8EURr7$JlRKQ86ܝٙi4`<Vsr*Sl#696`d=kn@;n!zapMSFDweƋx[;,y}Im7Gn c2ӕ[W BAXEWjl` _C,Vs_>T?'=1i9lXvmn q&_Nݯ!v7?zQ41'c*]_mTCjy-i`ϫ(_.a6ê00L^Qt3 ڕ0RadB7OA:D[ 鹻:eD.9|p mfG'H~&{3G˩c+)0L=_( Z(`,l jEY];3R9|gE]5y%ў6ꏮvq[cL V0>a~ `o5WXɆ)V}-Zp5N%͢GɈ L`Q%1c)%`EW*f2.MGTKɮZR ' kv`~`be,]-OIS*\~:' J7(}Z{M愤MSeN֧ LZZQZgG8Rc(Y}09ޚ2]+5v9);<{GE4G}/vvl& np)L/;#Մ8bK" n<RG*IDmRƥ cO#ѧC*ׄQR.RY4Z.QLYk/e)O.J`3RiFNT" iS#-OKH$Sy@r#Sr& 8tI6PƻО[;pv*֙\ xB ?%nدWn.460#Q;1x06\>rwxԣ,d:%[P^+CfS2IbU NTר cUQ6-KCH"F-'VG 2: H2˥ Mem M*i91*G- d3R)}]`SzxGpڥ"[+> ┈!Pe[:|1V2Ni k 'Yam-7\fN ʼnGwd Pu od*TWn'fp W̴80DY4\Jĩ,H܃w~~kBk<]ztǗ ~&!hw oLkv[ɉ:vC2VXi-?ʹi(cmQg$?5@ [~-+~KڂȓQhq6;ziH"OHÖk a"mJ*q1`+kgƬg:!ڴrVG+Tvڀq1 `łv/͡U2 WazTP.ar |oͩm۞FSG!zS/seG^1;| 'X',}8ƃt:8ku-\WF$'`ɨGWêGZK}O摐`.)jo-9J '!TaPg xkZ^9^,ҊQ2(4x3DH2" A?Xx/hLW;F|G|%OF3}XQrx6G>оn%X6<ɛ`}Wg'SKLqV|Ϛw>Wyk$w=Fe[Zj'Ea| Շ\idTtm@֞Qp1|o8Nf֒"W>aLICq`ޛCݙB!⬌]੆xc!& ) 9H! D-ୱ@L)?x6/yY鋪l?ޛ^u'zo-ҵ3чc5CP42(!6s y ޗ~NkEڿtjZrtU4 JAR8+Y׏&sTfW46|@Y\`jłLn9 |=?+;w?U3 <=0BqltFlpwg8@P]{es+x~QMTԪlG4.d=jX/ |gI_C.XrZhI>!^Tqt:{9ã?g;s6o{!sewәx{^ E :h~^J?>E^9R~)mf{e/Й ίҽ@e~ѕL,0>1~7b31#h8*Qr(6K5t.jni6yֈѝ R2PSo6`jfh%Ȅ-4ô"g؆'y:/Jmi8[⛢6$]ݑV^α{*S Wa- ٷW8ΖѪ.{CХC_׋8Ѡ-*uõ S݈ VTeU{[#)' m}tus¬(EH?MFdf6֖n%ȶ$-.ۣӽr{n*ˆ[ܻw| XbVy5/*Y@m;}Rv3uYI4WњU]u;w;c 7UUWѻ]rJQ 9s,tC5-g0QJr)ΕRb%ަ!ӂwy;&:sO"+CJoƸ4pLǴZ`a4aUxf8r􆻸aˤ8wmeg3Y(qĺ/2JyJI~͘`3>+7o3ڼnB*IBݑ^R!TK^UAdIR_I' t*e}/ݤZVY uzų]{Ҩ }1.mZ/ֈ,MYR\%.km7@;Bd,.~u*}Yiጕ5fYKcJqPB ~a,װzlLLz*X8Ն/ ]|QULx-*YZE2!q;E9*;`)z6*-$2i i~Y NRR $D$7XY+y[m\mĴk1NvNL9JK|y(DigĘ&L`#NU5KxO h,b +iӵ2o[cަB(VhQR~,cdq"XX rFz([YաyQ ڭmn-@'Q'G/-s/ <L.-%$ˡ\$t9&'cM|Su&Qsݡꛉ{n Ȓ+:"<:bRQĵ4!"N!.m Ƥ蔧HZF4H-bOqx6g 3frP>{iي} z̏#x0Nk[-lOG .7~vKq{_c{=`ۣS_Ӡ>`Z8JO;Yd`%'^H yW e8lo5O;Yp̡|.bFs6(d[<hlUxn_).*yˡ7#IOk`Sb53a)n6i6 }V7Q]]Œǐ0##"Ȇڍ>YcfFYϻP m@~_OknVN7wPVBzͧq~~{wrٰ/0[}nRu]xtv5#4T;#a"W} Do]Cx^k\Sk0 >.a0~eB-iP PqGE<1Ω+Թ:Ω+ع:Ωsꜚ$T1 Ω0 xiLۛHV>hw]!¡!iwu_lBƳZG@(/#nD >pB tA+>$v1%[@9B)ƅP> sEPSW$E-FC0m(rzÉ9tӎ:ZD0WM?88vpR^܀D%w^5׷VM*x.c'K.},s+7~\'ʃ|b}γ|ҷ(}(gs*&֖$ޱ,K2eXӥYDQpQQ^o|}.bf\Xós~0[CWx/Wx' '1NN3(;/28}Dpk_Jg9[aw nlT*J3ʓI*U:ggJV|$R 橎ãX`FHcҾm"mp: 1fv #?ѷbWB<z58D-za?ö3SrS`-bu7BːD:od>]T¾e+n|wQ=~[:r- <%(\d{Tb^ڬOA,e|][dM:{WkpG5"2&;b֑&ъ S:eFM^ IkGL̬4kCyQrquG^:dA^gr-`-'S~=ZB-]t Q .!h=gl} LsMI.>NiQ#Wқ8B"9E*.Ɛ8껐*$h0)h1%PbP+>ήvN.ޕf\H|VT{mּkpogzWL BXXQS QiXJ,2, I lpreCv6X6e6E~|mFF{Ez,x|jBn1Ö)~ sLVaXʊ"I<:ĭ0Y2a BBv9!Z59w />^|5K?IA-J q>10->$(Yrc H2qrPWNЧE=X $=F@B$dMKpA~.n$zQRpikHlMB5x.SF)s`K~wwUl?>Tӈ9_N-ݎ6V >;V[h@p毧\ߧ*|oI%Xpp}Xj̴ U3?r,?'4oY BÕ֔tݱ<;6pn𱋈QɎ!(2־%%8-a:jS0ȹ%b6#+_5f<-ۑٌk erǝd.9I׀t g 9$S2 Y$5eirv Jq|hp!BUauӔU X5F1<"Ǒ.Һf`:eOM9I2F-f@%3-'GZNeTF:( ]{oa[], zBNȉު̴%Fȩ^vˌm yO+:^L寅,MuE G@.ΛHZ؈ã!ry_"<eEpfZa%qT4s#84m::P}FJj=.F~WTs5V\"Ԡk.@]RKkônP=Ŭ)/䧯MlcʑE͇3s:G^\.QAR`dpHH3oSyѨ Q}*dy\B15We٨͞ #}Y_\[ @:;Ԫ}0*ClSj[}uH:뜠RFc[1{ Y笼VmvEݮ!nS^+L 7º&IVV](6u5;_[ё٭iyao"2bgU}'yLq2Oe|Z18?Ÿ"Yky=ɥ}qi7ο}تk5b۪EܦxNsGTVkbjEu.Co+ުУ̃my0m<:X@Z}KB1m"FWЫ1; (hPv? jVxP,[n .v-uiN1, ӌaY̔.MPm Df y_x2)N AHB 3BHK۲HͲH("bR,\8"h$4*K"n==^#iQ.g~S"8܏U$GX>N ]J$xpI*ţ!1E$OXk)4%aLΊ9i*'oD*r֚xM䍒xElgg$GoDi5R{aY pk lE lc5K&G<ޗbEE<'\ʰ4ǟPFǟ&h:%F-'CN(d~ZfŅm ۸N95,X@`c&wrj۩ _D7vblJj˜v},OӜufGOq1O'X2gf8g˺)-FW.", Ow8 I`&Ou;cO"7#sP ppܹ01St8[fƼKYLRo/_b>aFL+ o607yU :p!h`$Y|/y$ՔohI`^]lc#r q=E;XwK$N)t&L`t01)}!X4r1%'|zWl{TÅ΍ޒ&k\ kYk+1bhS&j3ѡR㙭vxSY-Ch9y_+Yxxge*x|#dL㆐q}x/L; kc*r;VrOf=KGOtY3&GHOposxOVPxpiUa14Q}Eh"?<0$**\${0ry0C&Hx}laG 4bqŒl}MΕ4bVk̈́Q LfEWnTGXz{.y5Nk`Lߐp3V.L}uY?y_1xHgś\P\d}q ;L=G<>ܡm1UMDLzSKE1L\_xbLh\QWp6%Ѯ"+Z6!?'4l NS+|\Eܥck "x4-Ay_Y!\!C)IU,a QY[Jvi{H[?G0|w1S|jK 翾|M=w|f@{?RRGu5wY8>`zvB8KވB̚r-\UQU*J]!BI7H r%w_y}e_&} cӬDXj`N1N^ǕrAZ 6x׌7cqiM۪9UD'IKP:`[a0 ~vA<= sjO@KAf_yඦpObj';ެBNIC vi[>KT27,K%"pi|x~V0~a]OlfZbDį=S,"qgZF"t^~,8p_ޅ6k#V86o@$ dl 1-Ed?8e$r/eǻYY:hܥHWVI?p1n% `&'sj35 R8;wRr~\cN%vg96ϻ2dAi,i g@9 У?>ԣ=xVx QtȱQJ~ejAQC\=ҁyP>ڦI)%ˏ\2_h:`oP;߄yRcBx3g=zxfy4J!beM>˳n:hCσr,ϺQ%1UyYc: ^81èj@[c"ONb oUDgOڈKvLT&q[;y;NɉMC&O"w!܅{RL% k"PNDDc4:?#PpO@**Ra1k*D#pYlv5N\{+lv9'ǴR6T S%`1zMe fQH\xmuK *gaMu*iQޓ:[8J9++n{hy%xC=N:v(كdtH6]=XxRFUK,NYj1&Fh, m% RF}"LK ;Y 9DJuĝrmq72!uOOiA`!KeYi-:2&St#`O{f)J雔u/dU#O=rFu(x[妓0h6ڂ=&;|6V*"  XȊ`/82f>VYܐxWak/{!IbĺO7i\obwajwC˛˺Y)'GZc FOɑ`P gc\C5pjVO˹ 'aa:a5ߧ?"T_;N>?_>b姹`i>do 0s޿=} k8q4ofħDZvw/YA 5׬b_t8B2g aU_8kD4=4w}ܔJ&xM2ˤV;Uk|Pyu)6mt #8u5 ǭ\.%6!`=e3jv ,<2zqeZ+NR$ gzV⽫A!6{*Xo$v~aۇX?loycs5n+DԷ k0vuzCpWh)'v_L֒EbӀ,˨z:M:[ϭ9KqzFuLJ 1lֹȽhM=5& o:|J*[Y3M^Nu;"E 4:XD2'Tfi%1"iH&ȋ+ ztN/փŚI2JU[/ oD8*)=b#+r{_EWfu4,$"RJZTǭ ,U|숝Ԥ~߆G!4B:_,+贏 bګ`Q`QLjT$OcWثi2[ h|>x4mk ŷ7ajg'>,*0*ʄp3 RN0X{)uv`+\IŅ?8dfvpDӇ GP k;w/!w-l6+sqkV~|;/o k;š)Y#n27Իg_33Mw g`C^j{t )jA4_ޡ0' yX F%a١xw쓀趮D?NKFpJ^vCwWfg vP=ԀyHjq>L hm}vPdU/?fKVg5)>∹ѡ#(FL)Y^$sKLbv:&aL] a_i*%b^dMNBp?XW鎃0vUe.'9z#t%*Ԉ -PBR &K؝Koy: RrLfAo?JW(Hh^'eJRk\!%A "Uk[MCvC,rZpo{%Kk}8|s<`tѹnj9bS㯃FG!pٚ#'/#2*l Ӏl ] %ri2phLzڞ)wTY5ڂy8#B{\ï׌us:SW@Eӓ>9Ϸݹ,KYƋtvT RQ a Y`6)WYBOtli47!87!Y|J (EbƤ`ΓIYVIXVsWL+}:k 3ZxYGN]I]I,g9MI>lsn:s*Px42 .+ͼEF8aw Ҽ@ۭ&2%MY Sߦ *,7XZx?7=Kzw(_d^ik&(Esr!oօ>c'KiyQnyƭI[^]|<_ΜrK:h|a"R?e=>p]Yh<S>H )~8ՖDq3,K% d*EDǐ!#= uر<Gy\r#'6_Ukyr}y>I^r!QX\e5*_+\eiƎEFߔjto[,!fq].!Vʆ*# -,FXO1/XZT/b޹v+|d!_%'U \PP&]2p\&:Y7^To׈' s)>'OҲrDzBNJ%i 6L+ l?'3d[r5ε75ҮwIH}ve{f߯(ZwC3K,Y)YI$9G4MOZXۊfP)d"Ic1LKnOO%Ґ1DsI\|aFo%1oaToU=i(-=̀.(A߮ 4&p]U"ڢ'ٯcGVؑsohGiMM,NgSetU=Ќ&LVjWa]?=w7i6^5V g'؅$3yJ T\/KY1n{LJ>r8IiK2L Z;"rH\(,.w:i8 hf d ۊ(Y?iKALC|gr5t#FZ1ƞ$k(+>2X,AtƚJ2OsP}́9* 7v1ㆠz-\zljk$H%/f7ˍ|:Kk'dh>MG!'r G<ˈÙ8alCsr".'xh*(!^k_{&?T.&#;pd0b2&sd\s[yZauG$OT=^A\Kr,`1Ba OB&TQ0^Tn ߬+:`LG~3qiJ*áTfe^頮H>NK`mTGfFb(H̕Do,qL}9=8z&|QZYբO:rcOҦQ(-Xha(I Z.3iH=/'N& O'4AĦLn4}V7Ψmݻ Aϔ,/ec2] !Zl 3%Nx(aSWǫy2!)bHzPzOgj3zj())PV)إ3vPayzW_z_d'.cʹ3qxruL?`3q+gg"sBssXzW_@,D_@&.3ƒ 9jh*8z&J^QEpJHY}=ɬ;w7ZF\L҉'c@$w3d I{ 4I|[&/qK~Fq4K>8s6@M͸s\`8sB,:צWL.8pE臭SDݫ rGQxz7'\Y;'ɉ '8")dܝsSYA\A" U!$\p$5 bQ#U^("θ dh D,`)Mv"UWY:d0JmDJޟ6͠=aUoN_һw˯hϓ*Pjd=SE_2N&:;w\!┣|Ӿ\!w=g\بC= &Gv?n3(eG׿N'&9"$s T+G]C.qLɮMY)5<3mN$2j|>~*dxپFھZ@[ׇ&6!ܞBhdV\|2l!҈q[pzFY8 U2g }F5tڍƚ40_)4bNL&q^$_Ѝ)ך`,Gi QixMsoS Rǹf^0lxVM _fFs MQLPxu}mE?L1\H/0&\JEYZDz&2NR]ـkb8X ^cgcC<,I#*Hdt`UɉV%饱&+p e7+c87p{d҂o24\^@:ViB͒Y^mZYi ta=T†d2+IU'&[afqYC#"sl!E6Tn`ӧhZC7~`1S,nTj J"BgY(aG غǁqo,+}Eja|NZ{MP™\zJ üYߐX2 3a(03ytActiܾ{4 9ܭ3K,# y#{ fi9.?_ND }׆="DD6ؗpf%G-m4ByjmAuN Fb d4"`XOc#4YT!ZMR:%a&MV[uVlؠr)ĈüU?"zb $ߊ(`^1^ t!j QqLt)8@*kې3O0\W ÿQ?^ُ؉K&4Yڧ'ZK|Hp<0mz?rh\ꜣ,<7/F~j*1<8OOyjlͨ몸o'6L{Y)zC0*?I‡bfOC;ĦcM{"|iQW OT u]j~&-"SEު.E5tB`]Oc{ۤ=(TMpR[qh=!8*YBZrx ~fQ5PExBsI$>? 3bE+/Ig̝hB`-yhnTghQU$#I <Ʉ!{4LCNJƠ9hE7.nTkӇeZm?|Fy SϦ~e Z Ν8*c%&&Z_b=W8|ohZVOD!2-Є)OjߩQBh܈r1w(xoHMӀQ0 e-ꭞ-Z@^)8_n JЋ*j}i7-2C?޿-†OǮ#0 uU;l>ykBYBi/o8ԈQwJo'H f]9\Z~"b+XPp75DP- rE U!i .3%-EFC#Ϝ‚ߢ.hi][\ٲ?`;r5v+m嚁 >k*|RR{|vOowo7QqJkZΛ߂bi V"8h&e"+tNswBaMIS)xEi '! ;;_E !dJ$ V $>_jdnP^~бe5tS7Ɩ"L *m l_Fbg@4X6 `Bmh|4$}m> ^Q{ClSeص/XF|Ǘ(Mc<3:M]4FYQ.[iI2n;I)űpL{,P]=n*, [iԠRf[_[_2B:}!.7Z6G("kg|^dUFk7⊺j(7mz|^u^1ƊjFE;w2%ɝPDoI.PMg)c1nS tkyr$ 2l HX*386) ^"}q[q}mMC31t@ IgWc16kFMvDu PR$h)&XODm0G#ţ*Ø`Yj}G+91Q3\菘Ru:mjx7hrh^7Ks|i%uszxMmy ih~^ĭ@D] a3G=i^Q7zxHkE!YbDQ dYfZ@$/w/]?qoG}sbﻳ:*Elcq}K[Bݔ˂'5to q,\Ys`gn(T┄+&g0o'0TT6 5`J*L_h;oEݼiaYKM%8Y,qp"3(OMeWP~Pb슮1ʊ:ڭgc"Nr \6D2~^WETcWt5qVٖI1B`KGB5>x+)ÌG ˶ʙҷg{{4j災4W\KOѺj:{ 1#o yKL.k@}Wkoc{`7jԼ; č ]3zEWoeEZȮs%VϥYd"F,c2 Tpw7_?=kOc#(.%*%,y_ZhD`9w[/ %ߦ9QV( M;"fE7!5 JEk~0PT+گ3Q*Y0B2z\ɅW ѭs,KK<{Ap Wvjߔ|Z` צN6|3w^t hGmL;q{D9ɘijtp ؾU'6@̼/rkg4k7%0G}[[w9ٰ e>TZ ^%0qy3O 3FQ@C0EP SPQJLQY[Jqrm6'织e#F6bCc9% às/={/i`+ >3!>0ǩ_>g*UtK[.d:nj $ ߠ# r? ڮ<.J9S0齺2)0Lp}lI%>~ݭBP30K%ldyrvϨ{CL#fYu)o?wg~jxiUJHP`Ӈe|N Qkgg#TqSVUOW]{N$ĭhy"¼iӬ~˞N;Di?Y7rI5hZ kx9xi`H/#Vi__L ׯ]~;ְ_*'\0'bnKGAtZo0^\O7.~6_WͯW336qD=`E˳?H`ft2pcCf)M:vϻs-arl/h-$B—eO?,W,o׫S_rBV>9#TTy*,;yf9 :YE+,\prY'ӼXTpT}+ ‡.Z \_NW16)Qߊq_~ <;vʪ=,XYRvyk˓)- wػGO 2θH^t[ \$Ⱥ 4 ~uPo`VH""_Ny.&)O!Uw<{8>!-\gcҚSIqW\\<%_20`FqN8z1 x)jkj~CĎeWbl?]+* yG՝c6A9T !Z*Dk{+@6s4W P\f&aÛG3zGepΊ$Rj-@+"Ƃ-)zd0ZDž' o_ԖLAE;WT (:<,%{:cܾ`ZYJ YÃsw{S򒤎F& 2N2/£28i=H,_is(_1&ur (eܓmo1<iduV+1!IqO<练WC_.Y-&b缠Rƭ xu(̶ U&6B:/nU'cFI݊61d,C3a,w]ᐗ55̚aQ{Dǘ$sP2͞dρC##)jtd{㻛NMvDu  hfīRMZ#<*cg8 ^e cW }sPukžtNL., ^-Uw<+Z|V} T#<*cʼ疣[VF ]6c~QotCT 6gDg$ mpMG lf¸ţÈu`m:&j[zt0of'!aQ~7:<ׅD{H~,YZGufGjɃV2 F Toi'9PwEIQ:|nN{*5}iq0nQ([1nZ%cS&lQ+_n/zYQяE%Ie՛ty_~Z^ ο'tTn9/=&9L۷=_xG=[qavP" u"g5Dቷ-<~7ze@|BrGep^+mqӫ"z'~TřQtml6F GWsGepz=TTj+2@~^VG"w~SeIأ"P"g9ByEIэe)PV6ڕ…svݤrs\\]Ye,XG+Bi,igVblFW1CN&{DvƮ ciM78gi&T>S.,]QCr Tª2Z}>?Z{siQ[V%9A CHDeNsdYR^{۶ \-0&mMz 67mndJ)'i~ϐE)M[cr[Uc3g0f>ML3GC8˻2/axolnszi2Y^CY>^ڍ<R~<mȸ{gYwq}\iiv lF rÆAQ]^NorH`sTp;;cqk#֎:;8dCbon'_ݮqB;&r<,yv@>gw#Xq/慙E ~~8~`;}5N_N7݂z:@\mY/3u!zE3K?{b'UYAu2MQc淗qQÇW!I)C%2ռ|}a sJՖj̧Ta`=BӉf+c2k˘ʝ/џVoM W湯)\b1b@*3 Y|)yUL͵_cSV?&6/B}_/o_j)E$%eC~ƏyhU匿R~NЗUO41yfeӇAoѾo (ٌ?EnѢzI[VuVԋ(ŤzAMc_. }fy:9=9aB#)ԄԕᢌxE\s)Sldf\nU]Y>ww4E &/7ܖ׃=pd2nVg҃53ES4``=4 13ñG ƚȫ{Qm;RNy3 GtH*ؕMܝavMQv=K?|2 !F#@Oo `Œzu2qn>,` 1`'7ILMĒPTy~V%t,8M~; :i𩖌@]AG'1Fe,MGO`"Od2@ LΑCł,Zb T(>Y&r̠3g@ʓ/A0BV8 8cgm@RFO,u*& 0w ȷṗֈٹ2g4G!Vdj"(8f kn3,#E]z3l 1 *CzQ$#YAÆL @M.P14~9/CҎ|}mڐDT=ERsQY5eC FIvQJw7ss,bmjC& ]y 7O ɳHzWl«s=m-ӼĿ]߽C^2Խ?奱]";AJD@6[h oz#فxDXq2q9~vgțk|rQT _CkbLy0SlqZNH,huz߈l2i,fG:4#XLjAm&IX8KJ3A0ɵ r B($<3 @\FbaJܞPn'}K Kv^kc ,w("͌ӵee1l]CX6zex: G`$m K/A>1yTȷ?S TROCrѱ::_Gc{Q[BQGdN2tfv*Mr4x\[-2$XK͋z}8M[ٸoEI#*Z5/Й⋪ӫJkM+ߠBRv0~9 6JʤV0=HV1 n}\1n0FޛnϵOnq:RZ˻ۋ<-u;_ aagJyMHVIIRjSf4.;˜<´ Rm7Nv ~g,l@h;"a~xՎpč)QB6ƒeiC#<&=\3DOʡcB)7j<8:"c`;X+\FB"#E9q\t3vch;iB;c)Gf?oj?8ѷ:0%::, (u1a`j-ȟ骠$6M(4I:xXd6G=PFXSJO9 )%F;9K#ɹ8'ᑠiExWukPA}.h3!u9\7\u=3ER1YHߺLN%_\CQ42`52M8򅙞 L󛄍@c}[+2,ujaf]s3M)n4"ErŽ6=&?:<{u\~sM7zLo==d5] lТ2kY x=ayG6\Zޡ.C]An?W&4HыP<]hB=FiPq9(|DbEm'&[ݽۓb-~X:5tNcTAA!+.ymL\hQ?#={Zf׿%Rʚ L,<-yF:H*0vVk3dXĻ}蚑ONOPC u!^WAڞq̒FfI{)LwԱ}vj'WjMخ|ӌ`$V{ځ#+*,mt=ɤJ_!;|s0@;y*aЎw9^8b{tfOq?{Ǎl@ EOXXYΌ.p4dIN#V YSlͭ7xJd(El pW ;<' }VʽGu6m4AE^|i0GN 3;F=Ψnx{"/_- '<}MwM(`7_^G.ZcR{p1 ;#b )|τ* qS1w2ϩS&26&Fx?3R1)QPM$JK~eNE]#bP1{[ Xb?~5fp|'p3V=u?ov>k 4ޫDH=gt} S68`W.-vG_q'\!,/_m5k{㢞.7 #I)Z-2J.S :}6977H8~ཱ4|L3%VAEk&ϐ0|{\FVעuS&\m1?֪i #Gf45F/$'9$㚪R^ Z Y1[(?-ax-mK ]Bd, nQlV689FUw>7Y{I Sٻ R4%m8~!mTQSZF}d #axW2jBUjME$k`]HsAI֠Xs+ bymE=te #;)h i8+,69TB~tÄ#axjo<,e#LgUiD* #a8| %'E*ęMjJ?Gd j>xVV]e/]j Cq^'$PebVT s51gx iVuJLX7 9 Z1Ur 3$ DF#hF!JG`' 7OƩ)z;G`9WQ n i J* m69Uϑ05rcPro$Y$YO&'B #;pEL2 !ͥ[p`涴R\?-ax :rp"-)h])FrRԜ0%r2pU.H+m4e3$ Iat \2xTgȆd/6\ ' #+Fۼ5W+p#'1 ,`|UL9Fnx>yFpVGI4~3j4BC$j^AQlL.pxs$|ch [#u1Np9Fnxu7UI {Zb]4U+?GNDbe4$x2pϐ0<* Jڐ{QB‰K\y"ϑ0RJU@X]cіwI]O'C7 NnqG#Q^Y$̐Ǫ!VoWk~]w-ˋ3^f{}Y:HU mDCP^*B@GEg j9L 9DX%n 3g7}->3WB'J"ORM9F.7U5XAA S(9/F5ZDj -Q)3$O!QF94+8MID^'+: 89FVr0prs#c茸՘*cS9F _xQ*VjTA -k?C9pBúeiqGBj?G ʑ4yc1EZMJq8~j4L n% MOkRYL.~" #axM4:eI BlWl"Pڌز9GFT&9#MJlx7@զ͑04)G.VXAN{"T\2cv}Jd@j_\Jc#- Hf9vK÷iFX+EMREm՚V\6a"#ax/R3;M+J[ cX 9s$7ۃɔdDZjf3"WĺƮR-S6?GٳɚDy.<\%TN #-ZAPe{b.,;Re #0<Ե\DPEhqLԢ≬n^JEhGa\t"Sr@$ܸmj ߯m^W|}qqE?-6>:h{˛;N]l_o;~s'$<s^!񖢿Q/7/#<݌{G757][K?:R,K|<:賣wѿώL+I+U,COyxYh Mf1eRd ̪rȜ>9u 5\Rk-DYSk0هJͩ%.]t.zom 8媟 vzvgkO.-c3YJ2/Uւ8*->=߼H|:/G2 4ۣ=~?GgZ >D#D-9 UKdZ=$ɕ bgA)6GW?}byd0hC@nun8SfϿ;zp^>^$r)/W7*"a7RRkn}52,k.O2 8Mi`|d\?MH.\M WVU)d!p듢,NfWI*#W,:@~D& М9dJIk? sx Di6g"IPs&%L多-Vr Z\C5UgSUG{D¼.7`yh<'(&3; DJh "]p[}ӣخ{ڮj}:ܨ'NCFhGt_j,9f4z⃍vՕ婸89>>f!*TmNlU[99Y4Jҕo7;;h.cQ RI)D"{K~`wx8wz{ZD O owFxw]?ݨ~t=`5;F|9mp}E z|{ԏˇ|E?ҴZ&^/ŝZ:>JB.e]tb/H $& DB,?[0izeY7Ռ4.TmsZĿ:Yiye|ZYBΡ})4J`|VK0JOՐ K .3Ҋ̴ٰ]0 #Öf, |.E,ܕ̬juMIRdo3Vl#MIu3o&R6Ԣ&]e]lKUj vmu}_RHj Ze}'=I;㽮~6_}zA. ^\ߵ4V'VW|ZXEq\mQ>l%& 0Y rx2ǐ]i9DZnս~cOP?O7}<tŬ7}ګb*V{חgE\ xNFiXs6覫)Jٱ*8.o4ٳgd+[έ<[۸vdF`W"O5>3f'k66 _\5@!K%LYr<ƙǎrqb4a?uWXK:;Χ -bs޲cҗ`*Ǿˉ^iײjz\/=ڵ-,;Og}o~碔MY#3W[lM6/v_eХ܈8Y|<^חnUʿN7o5^\_\uAվ7zo++}[}[*"M( |O(gW 6ۑ|iWF[+y)IhlE (, qjQ((O|mೱz)t2 ek&ǑȮK8RT7Q+41$gKP)t-(HͅMaAu`elHݐn/ӻǬӤ[XeW%f\iq?oP㷋˴1h'ON;;99^Z~>7{F(sQZ J=Ғ#~1ko 0xc)(tuOWUBg!N ,՝9_{U?(7l, nToϲ?jrB(IH jJ'MdX3Id<'sEՠ|A6{ѣsMrITgl8V30RSE6GS̄/M_ Rtyu^nT+'N?";޼^7G/_Qt% z;]PwᏍ:`Doo@ZEHs דք9 _.os t/%dAZϗZ8?,ޞ}}YWf|FR/ ["n̫i{%ׄԛd \Lׅ{W_~υ*"d-.~ ^^ox0r8:|C͵a!Y]|J*GcȿSh-%r?bf#)޿{2xpÏޅ6s9"i ri K? zabh>ח $?B~uXvo`.%oB棶SeBN~"R:Viub>!t}Ka ;A,ł$pAk,N%&@ڀie{g"[' |MGta9~i$"&IkdLI ={\+<sO[(s8(~j! U޹2J$F (Ea`Es}ݏ'88]t5Y3\d1\-䚹"et㠝t Lkylnt}lO,05  4q0w ^9 e0Q:qYBjB-Kr R#.滏)Y? "Pc1ޗW/CNQ,?/sUi'sT T %VÍH퍗,ٝUhS6тeYv h ,.r—4.ףUoe6gsZbjN+?vMHkB*TzDsyw&27R$ӀR3=K1ĥTL$AL(΄DC9* Y@gLf(Jn(m2>9#YR4ـ\O֣&]"5Gys:z!-lv%Zo /Y}+~g]ezL;tD8XOiWZ8U)'y܂'R-,-# gsSQwF0Z16JP %&2` ܙ!L(1*7F+B7Ċ3HxQKl 㫐ߝ K׷ū?}mnz ݨo{RoOm4.Ht9 ljmVä/њw . ݴL⾽~$ݟorF(a‚9]<@,*8.H)H9w%X#T&gÁO .1Ǐ W;6F\v]\,vq!-M"wuk,)ZzĺDT1q.@.& O5Zks 1%ᩋF]dGsˑoWo|w#dk'@޵K~`ρ^~$!UxՋbrmR F{;Tzrb~N1]&λw6s8>(F nu|VYֹvdjE5M8%T[MRDսԏԇžEH6~u"?ӝ:=-ppR r/vO6'ۨˊ;rrmzP+Ыa?šOf64#hU]tVI!B[dBUԔ 0`iH]JmC+-I![YgM5amL3HD2Q)\)Bxv,;BDz≭Owom 72 Z{n"O}"A RjpXt$]I44#'A`?G *[}ےdb3ȑjPIEX^22Y=KBARȉ5e9k;k޶4r"WƂx %2XYJ,xm N4x(Nm*hR)?Ukʇā&C #"$~kwpѕANG ?T@ZZ~N'y[8(^R(9]ʵV2iG0w!YO;HL1b:CUS/ѓ}mWfc]Ln+Z\yS_xy68aqѿC9_{U? )7l, To)~&JR7RILưF2T9"jP> 1f/ztNw_582r5 6ՠhϩ1x4eLX 24 ӺKwh1ˍjrEy^5\Gdvpz՛ˣ?}#?޼Dmqt}c]cA.VmFz3\5!mzסr݋C YPĵ% g__VUosozW!of]{(ǽgu]"nVXI&VJ ʤO~H=ܠ~ ]MUD[\>/]i Ň)~jōpww x',/FSwC&7 b msDD7d8vB|/oxrnɹ:o~?JCk3 rBd*t.D 4`u e3:1׳m1[<~jCP, Rz8ؠwg'\!R1ρ*^!ݡVv*5x"ķ#zAOßV.ap)OQOmB"wL,8 QJ5"VtMFb|#G+uEZӍ /Ga) A*-K{ETL8RK3[ʜ J-*,66Jun>2eV+!%SJpee!')B:<(iI' qC 4UɮřP 82Q1xΒj=kZwb9Oe"d)p8c%2Tu a"U'=h=$-\e)*b,oƾa۴8l OiWZ8A8|-:"E2RY݉)pLL09ug{ykt!~5)8^0#+p`J[j9 \41!קmC'c[$c89!?؏@[3BA;܁x\Οsr2&sGp{rc?ϊr9O3>9Od^}rv` dœ$g/Kr$g/h !V+U{WF zftMɺE6{Y$۪@ު${-Z+&V9)oDwwWڰ ^ؒ]6XHӂ攢ͭu k8O,o([3*vbh Id=Uz.zɂ|Í p5(JT5H]:*&WgqRIʊZ.dUΖUcw[1lzM`&i(<^UTE<ayH$P>.~)LQ& 7@|lSr^\cDTQ|a9a3-Muyo@Kl1}EלBL.&[Yݓmw tj%+Nꊎ7h ~RP/#Jg9ch>CsМ14g Cs| &R{CZ7dwy D(R!J --% a]?4D$bxi.RHQ: L1y(Rd;uE*LƄ?obRƁŘD]jW Jz X!YC`4#_.3qך&-|۳0x((fojvhʟ$otU6a=yr7Cz d\(wp)BSϘ8P=x:FDz4‚Sg[r7 74?W{\y"浧c3_ϦP POJ7e%Dξ|eR 2%3"7IJ1))gZ` @#W]2򕑜lՎ; S8QD)#De ak87w tkCJ'*pDBJAY&D4rfARH:ujdcJP!:IP-R<w=[5f̤ƨ-G\;,j)6Qᰋ0'޲ R s;r88 gRa52h]yP sF30*ugEC%DhVLZ"$"zIT208G"W A*!lh1u֝-?U:PcIƑNc$!:5, AXxNwe Kq_=,% |zL/[LPo!H#e8j7k(R4S 9P3O5;ZWR^+@  wD*8Chɕ ZYbޣu#8&az5j)6g:L|qbVWޕL{:\sir&ތQӰK0, h^zS^Y"HO0G |6mYDoWP&:XcaHdWc3߼쁣Pzx3a73Ma&Ϩh,?1Gwc>Bv'xQFЫIROlËƋrt3kW@"ɁMrcG(:t0%(^HpPgGv]*(; hk?x8X8T ruHhi GȌG S*aQe /djf/@xqfYn-_R!@^V -]LLJ{@F%7s{I=ѩr!K>] J:R"*̰4A*p0AP8@29QtXNYS쏴u+{)0T#22@\/AP"z 닭 K J^:0&fZ1B48Ad!lIiH-gЈw8%2Ɍ-#b)b,5 Ş;8.z46b'%5Xqdx7l-w_b>WT~DgO22'э2%Jh&4V'mu493aZ MPr8~2#*9\EW |_44~1E}{2_ ՅWWӏ/ΛJ3!]t| bwNMb> oUG3XLn? BZjmKnr]3d}3mfޅi`X>`Ť!DO]w N˵twAkuX˔TNR`c0#0pҫa] c,':Rx~YoP_.\o_~zޤ^}Ǐ՛w0Q|@}30j1l݅ꀦ-:p鿞GӲyTM+l·iv95.F!R-Tp@/^}y5{Nx:UH ^#euaGu6ޤpWW  NSDL/橘cb#}!]T~-9-_(~>yWu0q0>[~zɾZNz\2]J`y#t8;%!xogkG) 뛿˭ݥgDgX8xf8r5T߹Ѻ9˅@çrGF\PRy` ljKA1!C- ,#1 q%Q&Ӗ1ʌWYLJ$(u{ES+`#kP'Cʑ3x&Ļ=yA'*;B1R=3Ɠ+c̱r:cR R8v֖R2|%gS4a#bN>#r]K QR Ri]ח\,rS'W .l)he=V YiԦznҫSoV8Y R2UMWa76ʚP L,-yF:H*0FQnhVWv\:WBߺmAPK]T݆fx ,UHvc`Ư*oͨՄiLuGsG'GOvV> w>뫳׽ \ǚ+|Ff6GXoιq4[v蟪쎊֎ro%jioig?S0뇜 y|~o5 Aam}\_O8Vp3LKͯưB{u[/Jlî/]QioR% 6(S–I1AnmwJXKCзnaaw849Vȣ,2"6)Ρ Ns-ٜ`Y =]YەU$*HX^<(֪.MNFs.HRdY)_6PA=/B# d=z!ZU'Nqk}pMN1)ڧ={<;pK߳?[r=m`D%Rb*ROvvq7 ytPT=@ >.(VF ٱ²t]XچUqUM ve_ $vD\'!P'(z9T0k^b*Pr u,K޹&g N"gT,x [sT%:w5m0ǘg%6&fa+˜ѥhr)j\˄Nv.CZ, _of}`r{Z qsQǀ %C G )ʸ 5 ^6b&ȌlsH4]+a?!,+.x8MɔQ bbA$L411ۄuG.c=cp4Ԥ@r5rD.@I'Q&Y(r%akpZH#mH;i`"Ch!jUu.ϣ^Ɵ ~+ʾzZvZ*V/W s*&VVeUl28!}9MGϲ>jKnCʬydzI2#!d`dLLglp$ں:47DJ4JQ`FFECEJ,S2);=3Ԥ8JQgipuWT7ui=pE%gI.JK.iroo'{UIɵgE MX+4 ycL 5AɐEtD3ŀ.=D"BXaPܥ Q@RBI<쌗R%PTRЈhYYsZxLA )G ȠQ!Gr2Z &h7^eJ21ag:#$I/i(]ChCyk,SĜB7R1y@KTd|ٹlE.yZl\ՊQltTX6$#҆GhwRJ$X%1PίdqaIn{wΧj䮄8pa6zVq npM@;wz)26zSMyRzv|q:x>b6hkؠ=-ˉ}bҞRRIerZ,{Fף,L:4_HJ."LJV8M?O7Τ\I?sn.LK[ŷrIjmjZ/g|4cT!4 QPlm :+和Yk]>MrZ dE~jJsfu.}5|ѱ Wq2P{e9x@Z_H8*(2-uOFЊo΃k^9T-B慞 2G'WqhxVZr&lY=߭~p + ZNeg˲$ށŞL3ns;of8x\y܎JR( u8m/V:Ե\91(Z J쑘MJ|-?]oB%Hw嬯dzKQ5۔·%\HeD`tw2~*Lyo>bPc3\i${$d+'O[ygJPxc~5= Cʆ8dUs*1\Y,</qL;sf ? Āˮy=)~ju?4e2q_MK*͙@/Y]/9ߵYoQ3tBA(0D(7",X1E-ے+YAʈ1i1"rPJ pP+H1HpD_:k4&ގ>>#BgF5x=nryXys}hG_;X6D&:e6" ɹL[Gw.GRXcRP});Nm< ˏlcz}pr ]ХV؝7*-vrif"WFy@cU"mNPUUC 0[ym*Ņ'=@u!-:}SnT֤ UNֆ(#Ig5%j\~0uqcIXrΖ6i8c>.$d\`l0GCK|>&ln1N: ->:J2fI G&%bWVE%% z N=^N>ӻRrICx1$<A`A3*5^h 4yѸn1q/:`Œ[ܔ8oM(ٰi"e +4,ked%g0 S NxL4/z#7s֝Gvw=>u"t y~`AY`$k+9'#ϙ쒏˜5}<dlHљ<jGw2wuv<(G#;CN7^M$v~a6pn7̷ǎ6&*X@%U,x64s7t־n@yqWk؈u̔C!Ί(* "54Hvi2B!r=J@76w/"H:IhYyWi:翾?C͇?~xû\ow?: &54Q'T]UMͫWmgަ^+7Ի @\Ja1৻/oϗP|/ի e1c#~,yq_f̯k+$Y[F4\N=0y?Y? fF?2\^~Pe7~nFX|=u.ݾ^_~$AG zTݟ`NgojY 1@-=+e,d'%1#a[G{|a3^=Mhƺ&eI Xhil B2NJPm4839ZjoM1@omm1bqo$&&ELz .#$W j! !X *Y_ FW:Y:ĝęl$~ozZyi 7Bf#D57kCךߧ%0-ItϕFޑ#/K8B±}vޚ=nNl8{nځQgY.Vp&cŃalW~b&fYmg[w{+PǙR)4/a1kT/4h$%ټz& w!{Bb)2W:BOdv#-uI7dV\.3q/8iuiR2J#N]=]i&SYSrn [ML&X Y,tBOFWuqsuew$B&Wvu93ՠ4dJ?OHv$_6vnnGWo o':TTa?FO_e&r-B;@ǧŞL?{W؍!X`k(^ ɇvg.OYr$ݝ[|:,ْ%[|㖞HbJʮssw|Go6_\nmAd;N͠F= `'wvϳfýf`056{U=L#5 ͻ%A72~]lc^uU̾+t4Z&yn֍R蚕}dQk$@΍W/ɖFl&[klߵ隗F.Qx<@"!5ﹾk^xL2Qs7Y]Ysberzw>3ryAGZ޲ڵBtJiŴ4}4Ypjv].kvjZ/v].+.erY\.erY\.erY\.erY\.erY\.er.*}Vz!PG]j}PtqPP1@#3sP=*IaʼnR娀؛ߵo:\W:#|ח21}ꊥ׎Q笲/m~`+?7g7 8A|`jG(6Oq t/nG4 mku R{*Qf OE=B!6i%g9n?f|1\fD Ȉ >@!$0p f~OA9@;ڬݕ9EFe8TXH,r?w- GlL7O4v|^" 0^[8Jco5F$Ѱd֑{ř*WDUڹMT& CHQs縖^sN:DOG #օeJ)x 2xL = 2\@IBHk"tD,b`)C?eXfUߍp~:]ȅ8r+׀6>8m%"J^D1JYshS 8ԼrUsCKKFp,PTNw5%+"#X+ޚقPPA>+vd}Ld#b dm2,D:# \9,,WZ䛷FHb50#+HM0-s.rIkjcZJ^!چ*c$c'd\dx}m.lըyތ Beg ^pota?vg͗`XH wIA3F6|&'<P5]%5l2zPp]}Ho|i=!&ji.*`wItOZ#U!2.aMo(ްACL^p8\5CVu&(SdLQeN1q_ƅY6WM6ݜZH|*s6Wha.q5$NzT>`S&/->@Y.x.@.GPM*$ <#:j>1aJ|B#SF9xݹL?{9 egS9.-} O\L\cy1׼^f=;ΘQpHN\KJ"yFSo 1Kup/:*ؗ LAmzIwŋ]yO\ipl#a_>Ihc"1MPΡZHY"!(OMTW;HR*:\{:}#mat8t 46;b<]rG1~?MNg7{3 Sg%>|ݬӭ|S-il/GUO?4O >ݮ??̟~yxtϠgFq/f; >͂[}6bg0_oiKk>=o1GOtw xm/dhXs壇f(L43 g=i4i|ii+3f٬K[3g9 c飅w`c}w?wyvChgњJQS}}/GLTݜq_(ILԿ ~SOqFwml?/釟[=6H`ryf甿nCF7UHI޲Z;_[-rmntԬZaNg䙹]>jK$q6s}̼4{Q)Sh(cc歵>0>~PR3x/"Ő!e(T@)٭Ek4Acd$ҐTـ_xeqq!QvA Xl~;qh(aC6 B$"ѨPgtdGHHoV6 )/R4 :rB CPBq ҷKMgx6K&\eCr}@q.퍱r@(fIne,9ATZ=7\[kWpOĂ2 QzbbDO] `HAT44(= RD%8*dW3Kq& ȑjGd:*e{:"Ftkxa9+6$Qo(2L &Jg)!.s [FQ[OpJAT()_(XC@!˃Dr5޻@8Gʈ '#SQjZ~6~RNXṤ(^R(ZqQ=‹;#C4z`%)4O<`R@E'oU`ZF,p*J ߜKJ[$6 =+"o>(,|ă&8 D[M -`8.RtN6)OT?x[(_t㐷JH: i6w #M*kDNu\&$439f ?CuC3 S޹k%Z3+a`E>*y)B{h$RK#{? 3=W~/jPtسI( x]I1R}^kQ#ȯ(-GpҊqFO&ފh_w^o*r/&kRBrhaQuH2Bd 6X[J؄RXDދqutmC}g;uοya60ce *!Z$$&D%tjeߑ%!N SјR@(yT(L!/YdK(u!@<"ۨ-!aB,Uƪ#3" @85I*ʃ ༆(G=Y[E3硄ZVIV%KI뵄u L"ٳ~+/֓3J&v<7dɇL_$O_Kskqx,ou`iZ昫ZBCoclw0Gk%*H˴ aNxҍӄ7.Eu3Q]+ C^:enq :{w@ZuXvˋ>|il}C#otGg.&c \JB=krb7\ A1%)bQrk}d;Tl zM eNGNoi1rA aB.9\pBxMba&ܲ.' }#pZҝl^!vh q^SN6#Ǽ[ﭽ;09vi"OHVm>nf='c&Z‚t>SSrY>v(bkpuz9yݑ8ĥ:;\ kh}b/~ԅwroeto 7},`#ҺD(Hkm@(UcM F{UZo&n2ۻ~r AJS )ڜhsC'ÁDVHu mm"ǭ~f~Kn粴٬ Cc?>͟Yz'N't EK52J1hGo%NR^rB<)rh9OYոi]oskm#GJc.m <7I ^i뢴O~z$lr( nlVWbU7$/1Gn2[R h5s cRSδc՟1ɽlj-bL 5n8xT*D0"2OMD8H1pQQ!!$ZNEa-uTSupEdX[f[:R)3> .rvcD#usͥ {C^ '?#|aI̽APIRQuV,eD%QAJޑCF 6Pd4,gنspoo@ZF]bp t HxCu"j AXx B3ge2ed@N_>g | GzE~Q Е^x@^A-7;I =~b] RR)븐.pGNq箂Tτś1X= )K5R3|%gS4as[Bo2Dh<1S^ os3?ޘL1bP0vٽH sVk/J+E%K>8 ZL6ܳZP-f }P%M]}IC Ϊj0:]]S3k:0VZ峇-ӗ?xLݰ Z]SBu-jtZ UZ q C`R f 'vaVPC&׌DLTLk88 RTHwHx`r 8M(rneZGEd%w9lVJHH!$i~G"l׫TGE}`M\cYf37FG!CY2L.L{W$E4 /WgҤr-{TYZbjjQ^ai8/EwcpL5(:o;cƌL"^ˈiDk45[¯%9']$33CR5jxTs`.0 ,- T)$-VJ\>Tc|j)ʧ %ZW>RL@SNq飶ěU`̙ Yd("]JJeml80f5:v_'zCmp05`U42aD%u yI"R:"*fuhW$a iU3L1(h L> 'H rXy_T&K K"cHƴ (QHLd$&A:łKsg H吱e"iE"wqJtZ]:_]/;kyݿ[vO֪([%(58J R(58J R(Fq2辂Xm-P ΊZ>kUʍO:e"B=aQ}lИ^m " &ַ`\[k h6Y;>E^XoIp+Փ7m_˞%`nדI ;8eRhBaM/޻Oi |iw_o儉BU-f|f 㾼=[{8ib{9XL9@?{&]Wt"<D(O>'6R1erD38x3O|1ֺ'(#1ĠE)3\蔸ݓ.e+}٤?hc)*{]̯/= zK(N.}pyq3wO9g(X=SQnZx7SHyO$(!V^i?g9h:K~2𗞹?8MbWW闎OHd9t3M>7L2b_KƗGӁU3\Vgi1ur9x$>3:̓sI@s{M@7A)hpW[ZQ% ZV͋,YydcO#hp%J7bi1g;Jaظ,Պ/. 21w{9Xjn{k f0th/ϳ+rbE2Hf"5˺Hh\YP ==vVXawHy1 Ǹe=շV+$tS6Q I=rҒNȁN z,RnsH@915.; o&Iw 8^քhӮaWx?n a֥q hz87QjU< xJ+(|ZT^2'Ԯr3XQf ej bRNVl9Yqp8NH< n=.|qm UV@JA8N-,(NEy1YpT"m !B ~>lJa^iGjd9HysQ 9qv#Hi _{5HYeG>iҍQ厸9[`7r>lFJmv:ثp$cAT"7:Z!o/7e39 G#"]pTqcDL/#8ͅ RJ+cX~6(pJjR/A6+8EyG&$gkE7bq)ގsc1-vNˆ+)u`YQSX)=1A0ɵ r B($<3@cLWL#,8ZIAg7{z&<"A3_-":]_j1SvUކj8JZWwW&p_Rw| FV ܮ}_NOi[5_fior"bhc]ccbv5wg5m}>{c0zzK/LYe›M2;@w1!=VFH-~apN9LM9 5& Nh1p]O}t_)`D hT_ N0 G%&zqy٣m8fY[qli$Wjno\W nW:O7=?A՝[P&Ekc11)ZJ RE%Dk9{d tE톷c\khf1gyH6RCafXT!ʓH&)=T[Ɲ¾+aFF)de| ^x#苭 KS(?ڸϥkń hUIi`Dy˙A'4"ApN%E Ru wK GԳsAaqQ.II VTv |_ wSZ]AUb7O˱?p>|[fBcR䞹.?͊i-4i : W7_]n>ޫ!L~:2~L}[7Я+?Nf_\,+3U;A˫IS?˒jF/ |qeC!ՓP=1]ݐn4nޅIC3jyf0bfu>>]egu]v>^:NqƚcэKM}]b< W3o6ЎForW|wwo?O>b>^w?0xY䓇_eyT]+r>l|~#-Tp+@~~m]Ez1/vfµ{9H<-׍'0lϭvb 4/$30j?6]6҇=k{1dYV~K?,ʅꟳ^>LcM?柋}߆7^%?K&9Ox?_ݐ &d~_OZ Ѷx?y eؚK[}ȇ3^zsό&Cn߶\)mu@`4Val崏 d` ljKG$+s- nKt]T7;wf^/eB8mSeruI݂Q iW4E$-CYv<};C+ tdg_;łAڙW#E(]~_ ;btgC-B' $Lx2qm0/@s6lUrD42KjYKp_gӍk$5QmGs29Ȱ, x - `A$J&I^vlazy4 2e{$WUWo0uX,ʒ|p8 ِC8ˌvQ+gCO}Q=q?J__vH:,tu^ ;.G_G?  OcOm!gE9ᢓ>A(j TEm$J ^بL2w:}?ZqT3iFݷ͐oR,|<{+ V<| 7{l<笨pjj7L=O}{wnVtcE]ڤZ]]w*M77?T*[h,&Jڮ{Jj>iҳӮ#h6yaz~ޮoJS߂oov'sNaIbI%C/W_X"hyI.ݴL,ܴRWi盿2=-#3L.gȢh( š pӇTe,`-;8.6I– J >4RMZAh;0)IBTɻSu5`i˷ݜ.#hIq>LfÇ*d_X iML $u$u$u$u iɔ" 5=N`ȄVŜ!JlO%:1ޛ6T[g fXEmݙ֜Hpv qՖO>ׯj %[\6:Od;k 5lzSˆ /W)lX߱aNFZ˪0zw#-Vog NvF[&&I(PI&W&(DrIƄ~GJ^uc,Pb7A)SB@- #k42s3ډ)QWÍFqp).kAyPE 4lC8\9"N. 19kbhF0 Y0B8#@/a)ՠ(c*T vbsRN#0)w;{v}xt2&Ԑ1|H#*{.r짱.Fq\mJs,2CmƄ`M1*-)Ѹuf+eU/5B(VIWu1ș&!rl_ʴgsm;DƝL؛'F"faCPDI5iv3C`o$}RlRIJArInǨfT϶ߜrv0ivѣD v5(Ѥ,Mh&j*gmVf,8V«T.UŞTAuDrl9[7^è$g>&e#&);\bVdHH- h)_?U[C''KhŋSc֩uiEI1v TZ[SWpH KMO7c )CJ_O,') UKHeTG!j nu2@=)cS㛫n'1TĞӁ{^R!nf+4y`id<&m'kHx&nOQuF3sGr'&#VF/aL4x_9n9dy T;Sxy" OPm:].t@킖j%% oB_rCd/ ][U+E6B]{XZ`@(YjVȳ f"\*5/gЎ%t4e̸c]5[DՁB8&qFQ(}Tus[̭[itt t-BR E2@=eg5XGRtf'DeiD@ڇDA>(I QzTT5Vd{ E֦AA㮰-zm(.Og{o~v1TTj+ܦ3_j>D\ц~=L>ZȳsAz",`\~v]v>4Mm!gE9ᢓ>A(%@-R$C1ڻYZt{P7οu'$XZQ $gsDaK(S)GFr9$ ۞^ owKs/ķ&I"ˍUsFLZ IlQroW;v5U=f|zԏ)ב}+ =Eo}h<}B412X{EOES.8[R*1 KEKZ!#6U 9R ؜4:H jaߊ);jɇ}`zoc?Oɀ.\_~:`et^a!)cTZU%@'R BXr=0X7$a J`Tpٲ0,Ȩ,|!T0 bLm-$Ɇ c!N a,0@Zbɂ*Ub l1U`uHE2byAIۧA(6Or{U+ݣߠ>ȊE\řlq` S-}?9-]O!>ׯEt/Ow!Z<\#1D] ѽo!YIZn,CORZ~m]I-9Y8ˋR>6 ئo>] #/ْ^J}!k.{$6,cq\AZсh9qFޗUm矝,(p|✭Sv YۧU$J#7`ɑϡ(2`P2DC۟9,ۣ2Cwf38&'Lջ{֊MEJCI@.Kʌ"v< (U:=%dG)æQDZ4KVID.j٤[6=4Y G6[ΖߜϏj(r^WrmkUyÆշxeUhWٻFr$Wyf ^`z13$mT۲h`Seٲe;mIvaeR<2㋃cs'qkc8gra݄eA1yEg2XW/ss&fd jIN'N[0* B|:\1ԅN<+&RfųЅʵYCPBLM$cD@VV݊]Q1c9O6_iXzE@AF`Xl uw"DHTK )Ar'zj*Ѵ JA#X%FYyliAa)*[#U? [ 'B?~{Q&Cq)CaȳƅX$9D#)vşR2fx#>iz)@|,>a_^}>Q%ֳWU+r3#_yoُΙWӏ"n(pŚu?@x~ݪ_z% e4<:>K? kYـ'=wa|<~BlƑ\?Mèxqi/y)'-OJx}c?nZwGUQ4Ƶzri"Y6L>R NkJ&=yS7tj&~u6nhxr_~ۏ_~?enG~g8)A2CCkۇF!\o3r]bҐ"?{;N? T /ZSzc 0zi6skՓEgep@״@ F2 4Rе#YeY~+?M(.k>ʇy'\o~p|fktx/9i8c }bP!g/iȬnZ40+ 4|}iE?.+/sjDg|Wp.ːqjߴc=>{Pкv:. }![[SE`L s)K%ycQTkKzk멷-i>n]yk8oB#)~;9:24%,`j `C6Cm*(Ľ0vTQ+}ψO@ٵkhzkڢG5zUG();^k6_NYX9hVFeoRmႲ!,+"EX5 r_jil͒2nդGܮ&1O<+l.-Eh?? T}7V2O32j)r ‡`cs!?xePs݀̚ڐlӕ}ZKUo{fGƊGo(/,yNzi|A Qd(}P2ԬMP[D}SiPTt/} CG=V6}4:]!lM:/>AʌRX'*:!x!ޣ%9vV]Tuq,w>J-yPyZb<8vUd~erj0m~-oenzwxua϶6WZJqw;4R\UͥL7ˤKyeUXf_Uq4S7-Ize|F*hxǶm!{UoG^6_oQ!6QTƛbUA̚6;={]EDjԐJFFKplMLh0i(lYȸHHAVђ$EaQI8@F%RNJvt(0:}Oc^ҡBH)wRY,d2R޹`0X@RFHѩOwWU}BzHqfir:y I9E)u(pFC]d|Z"$@ս2#,|IĜ*t%UN$]t75K#~ņ4)ԜE@9 2hĬ?DRPkcl0vg;=AcFZ:^oh;h'i4]BҎY-yN/a16l|ׄHY^&ȷu¹o))؋ Y ف(zɛ(]*I؟vsosۨ|®KNvWmiHBi-K O*M*v5ka46XəPY!F-[{yXDtJU+95J#&!'Qu@ GD**FǿYW/z5j:q97.)_N;͔ અ%>]Xť׋^ۑ7Pʰ-^z^ہR 24"8 TSS/ RϢTa˲gyy lS]!gkbS;9RD!Ffk*"%%Pb:xvVvp^rKv^y⟶vg x+2~16Чܞ\7'*C7w]{W< Rj:nuں{~-.RMdo~mܲ͜Z{Ʒw^9z7{SŚɸlkм[:W[ ૕U$M3ِlm`|ZvSr}[໱D۹xMة5笓C N5(ju 2F3{g=^S}>І@ZQ()TҐtFB4llfSSkDQ"^.:̔l0lnj r*E9B9]Mv#}rW+|y}, Wx)LGtZyy}GU()z!`A} Jh$m7 -I*"fTPbY*I#36՜V?gg (/@M2o^&ʖD&;k)Kyr:Y%h^H&,F,f|=bZqtYK~} 3@( xb :yu*0OК֮I6iBdzU3UFIOyF5mFFY0!$&34@8f,9=FKѺa{Uۨ3x&66b-kP\fD?Bu$ lR1cmjU=_t_|1En,}@"6Oؚ5/ZJ6cR =h#7i&KgP/U$!w6ّdPT ɨt3blX]P h 4^A0}P]+~,OFg}~+сG J{jG*Z#GІ4lYi(?+V|&g"3OS_~k>/-s8fsAqeQ)sק0 ~};ih6DlCwLH⛟U.?7oN@CNS M%\h+YJ >ڜ#%"zwzA8QZEuj#&I")mzic?RLJMgÍ6Op"#e{ˠXgE8XrL[t%5-b*Z* иT|T)䄥@p9Yiu@(? \Dy-u)iyݪ7e)AFpzYLxUI ҪmYS>9Z/ |o "hx^>#c]F+,LkL:;Sʂ ɲeZ2$AgB!c3˜G)u j6/IRS$lNg$buVI  ؆-i=z=&aAeE̽wؑ5Bv,s8S9^Y>Ox _.'6؛j^5D){]^{4Q8QҸeh+lj(X&e&d,XrD9{6X<ʁ)Sנl[@:KL’lv6-Nct%!x k,_'qub^L=Y)5xf%SG,=EWЛ䍇5UB($`52 0ɘL %}쥴b3ߐe1%f%Jf;AUHtxtnGz3qf^6$.,^pm- mٮ-Yڛ-ɗJEdujbkS](tKiގ|vg;^Nd4bP?5Qekzn +Ya88? ^ rC G3o렏b3 v7'='6gڦrlytv:5wlYe=nz%n-.dR³9 CI+kbD4iU@ Td כKշezourTZ޵q$2?c3R;Hr8.3is- )9ݯzcj(QqǯGUlc|:vq6e\ÙcpbYW'XS3S,COܢeB]Wj'm"7lXwdOڶxzi} IʭI[7W տ}~ +:fT~CG{KU巷U_~g3 `cX^P➇T]o,Pvre1obJd+wI= hi_Pi}}튾uh%X~ip<)5NǼ=9Nz~|y)8C!b1}p\lDScIC^h1Z›5L~o灇-0Eh^o ^V=l=|b)ۡCÄSoo~seQjsCT6J!۠LnL )&C][.~gEeC q2ǩe2W$2ୄx :ǐ$*8ʝ7\ 3f4 Q=x:FYdr NVF?ciV{E۽}:n'4<>SLZ0|'D@PDz9L/@K0N9LN9 9& Nh1vD??n;o^7kiuCƾwK]?h7B2_[-6nQ38QxN>~D뿚ᇧ'WK9Jh&4V)tY~\>XkM ާӷ ?7xRϩz;+_O+p\>x}YS #1\Oq<{T4x2p7➇sP[OB.ں!h n^̇}XICL7B}.FW&ekmd[s^gV:@>L]jli=8v mhTMZ6637|_~x^߯_|ūo1Qorw/@ V*t:y!|hzog?~AײkB}κZ.>*`0Fn Mcr4s.M}k8UHsQRI\}/ +,$!@ }BTzZ:+FTtE/@M9J'URoU/o_ 6Xu>_5F's$緺f'gFY: )_9! 7Wژ/?{zk5[pPM:љL@\}~sU:~vPhh-OGpk1[[NȈ P8Ovګ`QkI-Vt `mݷۢVpy>uq03/2!QfR؝`R`+" CCThb$;$v #GӭR+NRkp@P5`gK.ɸ0z,W@S>.( &*)l)\!GYI*ua.O͆U%Aw0_Qt . ^IeH7oKOOxC_P" X]fbyvw'U%how@#IрmP9  球as!i0FI$2"ˉ\"LFC|6Eb[G ~݆_o]Ɂ%ǴuvԹp^oOJxwȇ='ߘ@+)u`YQ)AϬ6 m8/ )фD歼O$ob5iI|r`{%Lq WB훋sx|R3.z`rd$*ט)MbL;eln1yZkyNJ x2|p-= ˖d>\;EJ8E)2($p(#XfwP܀aa4MpxaR[ERcibC ކخ:]NG@|XʎPŌ*-1B8 E4\yk#S1alx6\V/mrHBrNJs s%rkeuiAAb UG 4 E^#h#̮WY kdcaKC-t"|lH%1QMKc5N}rg:\_6-ӎӆ {F2PJ% (R%4R[New:R|WUClgi槠(ElۤӳQu9fx.M3|}[?']yK+ :2}SBĎwyEjs+l͑K.e6j|heDKOt_<z8c'5{SJh<\ixQ՚r{Jo}QǷ/;? 8yOoFcgmI 9qkY"MyK_f``,3yl.}( [:TT{Hڇ.MX* !O&ŵ݌Xѕk+Pץ= v-h# ,^B>v -2jgw?=]IP$tҌ0'S̷}E6@S.=!5stnzƔ7+*sB^;^CK$axddÏ0oSa:+&# *Am60@jh)Lqyڈ&e~OGŁ"|*+CN'/6˪PiVo?<:@uKA?$jX"oqe7 ^T2GJ-Me5&wMzg;ٖwY_=l;r ;k͑$ Oa2Na_0\fW{v-nqE4[˙z"5_G߿<:=ẻMp4*Ds $>藛o7tӊÓ`O0sw:be9*a~q y/r%Gb_Bah-61-rp9L+ả#ĝ(Py%/Ě?{V81E6za2ymm˒WS<Ih[NvCᩯXEV}:)]ѷWQ`ʆ@j4I\(BYV\^d[ Hۘ )MV Ah<?!i'"sJx/V1ys,k*5s$Ʊa6!(hȆ ;R<Eǐp&ft+\] TL-M iM˘A9JFS͎,[Ǻr{vpVіi'=a{Xt6/=u:h?Wwt!=BP4ZWkj_5ƯZWkO_5~Ưְ7P5ƯZWkj_yZzZpEH՚ZWkj_񫝐O(nr}Vps3\(T:v+bp9lnSd=1yY§5O(kNWʚ;2EM%0&'m 7f`֋؜!glhNy`VB,,r2C âRr̉` S&΁[õ_G|~}99?%m<ߒ;=t{b}ޝҾ=]gû].>it}`!w߰.}qvw]z.ޒs]tjQKnzw4ŞNmvq=}ԲZ[v~M 9VP3\ݝ"ṗl<)|]鞎Njٴ[>qwwdClUsf]<-y`5|7 u(Qȭ/R|\=6'4T xwʛRNjELB^+ŬbV*uT0iL$ G-V̌0$V J="A|pGeRLoe/DiAFi Iy+&TRj. f\|}n8;J)aVFxh"KF#-$I$Ex%EH*<^8I[UVlLR >bd#P2deU@+Y]JOL$/ D@ (RыVыbGuPCNA&>6.SgRH^&Q2iw1Ĥ$+IHJ$i\jV\NZgOw\])*ad a!0DXN9j$h(d,hm:L-Fc8) r}I;#t9 ޶V{{tƙl-h6(R'+LbwEqY&S3,Dgnk %(itV%!aIW'0 gWmb&2}fJb5q믡vh>?a+l2:js9+ΧXPa Al)e_G*Dl:cɦt8NӾ t<(t$71=!4o3^@5q#A?+q)Ckk_sW1Ūt8h EU=<Σ*8őhO9#FW'dv~^ =?g?Z*/?40Zoi !@nr{2O?><חMz{d;Ժtnw|۸Jۧ lZJϗhennN7T&3K< &(Ia!p0E)WGQ7eoa?-( O7,Q9[`CM? M_;u][uwҼz_̡y|~{9$-2/jvN*~`4w${Ą)?%j(nǓWSm:y[zwHzR_(^cxq? *#c!Q9om]@`VV#H.Z18zTcntT}OVPK!"ZZk+$:WOSj2NbwJd)VlU#jX'J-C (L6 (KC`fq:S Qe#u\H}cbYP|\7YeE8fЗL{kXS 6/8se{;S?UXuHWG X_(` Hq^MG>%rru>& gRU:5ht*)b.:lfm^g:qq|Chn)hN7Y+B!49#Mf9燑S/JyZoYJϹ;dycr:@1YXf1eHчR)HA2'sNz)W 9:73}s):+XԂ7oQ<ϯFVy?嶓Y˳TeZ;d2&Axlv EUh,rjOMYrAVC.1ٜL9g2R-c5q[^ʋjX[h*BlnKWes zu/pz@A'bX*-VfE'dBbѤhT0*p ȟ̺,CLY[#&f&²6c)QQ8DnW2E@`JȢ(l4R։Eϖ/ƾ"9y\kԊ+cGSk;\*'Ng]_ߢ+CX!.sx-wzW4nô(y^.}ߛQQ7&|N8k2rla)?r6:g },;7S~2+gZ+_=Fes%\?<9Dt6rkb4MzU G?+Ɔd~NC÷qb6=.G ^i(b6\ozqsx Uw<]w]kVrcMy,]S|=QYi=ohZ??Բ҉^u6.#_??|?û?5 #""_]Yo#9+>Mc7<" ̠{v/3=;h,xݒo0uXu٦,%))&dD|Ypxpmuk֚`IO=}[>') #˸=i_?~8r9Ց»?UK,)K ~'ti͈?`R{zyA>;r♫YM#ݑ+*[Cq fm+T@L^k@z.Q!e $؜:kOmb!ພ|ă)Wdc |"d_S1)e :HCN Vdp:t4O bQ-=)JgD' hy4#CWϨ5jp:#tN(k>M")e.:Б0.h;#.;LCz!q `lUth #^[}k xr _j&A-LzU'NN,S憅w"w?wTָHi]dB&-VBǜ 8Wp$ްƲ:or,6!9zN*s̟7@m_mrU43]Sc #7Hƒ7LӫQ^h6yOGJtEzo.ZA-j.K(]tqCg):꿟,=:azJ(G1yƐ0I'd%HiDE''AY]ңCh7jXQwl9S(\>zX\k|~g7ܮ,RM?5OrM~6ߏ>7losI;q]$~ 7Ifft7v2cYww7^n l@xi[v{H 8LLh7||C]TqHu '+O>CQ.U/HC*tL$f(,:p4>cW40.%Dm{QVʠ)[pW"$H-r>8l|2]jl뛗F+QZL,yBH:TjEG ;!Y΀@ko~kҡu1h0h+yy"?\ j< Kk~盛Z7>'9wɌ=FHQo/kXn'WGJ$m@9PcEFȚ^p{(w{;o{wn;?ke,[/IVc_+ /}ɯmkiM6(捒.Ř2N0Q^|}IN8uEt yZ3tϗNLŢ) ekst XY1BrR砡`EM5TZoDA HS2HGD,X+5LAp6BZiwSޙ7 ﮷ALy߽Q ~Z9¢3͡Ϭ;Eg :@K\q.0O&/ɻ,uJxBȓn1%1tXru PJ!Y*%CCe c%17EV"A9:C`Y"RrH]%x`lP1l6=)^?k_C_bgEs1޿v'ouhӫ>%ݵ@h{׵4X* ѹuǭ篭Yt;:]Q2l;CnqEq6zm%FfWZn--nw[}}ո y;:N>UYMrqȕ⡵ܚ_4lC:mjPS[tZ{J9^st#9SsdEZ[:FHJiu2tV5DW14BPRJ)""29yDEW D^F l4d-1L مL ᔲ*j uD]V+=p$܀e7Zކ <#je=;BQG W4>訝&j+6Q" ےp $}3=hʼna֩»߻CZ3`t0E)DNhjhГ=#ǟS]T`~"3Fe #aYfL ,3 )dG2hmI&k'd!*GLVaX1Hd+5KQgmRd8%f;/÷JEVZS!f[oYs7yB<;FR=qZJV<CR ǁ<ڑ-NlfTQǛ2~Vh7&Kg$!ap69 8ẻ+zT8dQ: -R,BxE#ǩwL# dH-2ٙZ>r<>^Y. tTGTF&Rm6 v,xj:L=:lmF;Yr>?t+$Q$$$1();SdPw!L! 2qDNFGufS8`%N X|>M]{vYq2 W^2l5n݅'3z7FFRKZƹC&Y@LEr i64L7^M?2;i܍jnX 82 LF,eLHk>WR'Ǟ ShxE!NU ݅A4YX9XlhZ4`P:g/ŷ}0s kxRn?2ma {B B{z=D)b㈆!~gRe-RhuCiK[tuօL R IRBI1 Oѳ*=ڗ=ZhHDJK "8Vr>:H(-|>X#g?{ jIOS%|_߶Ǫ-^ g`8/ ' ݵv6 4ִσ~نG#'~[+,{^zͷ+Ju W6]1@#AL]s/|?}hG/.Q\Cԉm1afb5]t"B,%P:w%!msi& ☿^IYf~o%y_V{呹%i(/! Z5匆lI = JE?{Ƒl~0p>ۛH3~JW"ȡD-ǯ(C}NB2N AR-lB!շdm8%zr/7Uʞe,'gƗً:؞\ty7OAmw8eE9M31βDRy NBPxmc"^>lg˃60QT;T h2y2SKQBރĺD -î/ xhJiB(e򠀥H.8!(Q{@VQ h<ֱ!P8pZCrڃ0ņ$ hU=턣}l5/[DEp{ ?fXæ>YؐGV{U8vF)q1%HK#Qz!=BsL¨JEI< R`Ir1{m9X5{έkpS,!r&G6iOPٴ6I|y?R HSR(6"a j@ 0pRPGHЁ $ Km`u4#F0(7kELE cŎEJ؞9lqW8uu]]{US3(o)&f]Q,{,eݲ_Uі?8HTNڛhk.HD;2&FĤNNzR#ge'he u:6xx>]Ve'Fa <2ƞsUP}YtAZ t ̈́RS#E8(ƃC̪"0 վCK`HsofQ#U?|ry t$ugO n}W7CnUW!6Öq,f 1lg=x+#{dWj罎=fX }xѾʞVÉ^&k*&~<켙\3$ǿۗ_2~?~g/_L}?߾J ۬'CP{{ =pzU_]c]kB|o/;]eJȂ n s?W-qӳ&8|D`73殽j=mۄ*na#o0K(s^`P&5> JLtiډ~KE""d<]T~ ?CoӊKM{y3ُy57]ƋwW).SPwWxT~wC& xجXh]3tU_ݿMQ\6}~se몡Zۅ[![[ބGm9!]@ :Z"R:\29)cںkL-|9FőA$$I8H`Y\hs+ $x#A;tTzP >P̟S]<4j59"\U'JQ9Am |0'oG Sni6{R'S6l]URTrjUKo?/J 1 Zk@\S;l-)a Fr6bUDY&Պ t *'7BJ*!,rTHG"Jޒ~޷[eȪjXyVs.tϾjcGqwg_ЀL9cg,^`b>(c.a 0 ¼SM;Hj`s%Tu|[L{2t<}1̲t_M_Z8A4}Ir>:RHuvN8S'&rT,A2t;_fn}v9џ8b+=6JPt,54rg<băCx +e^rSnr֬Xg VʻX}n(ׇi1*#[IU-ߍ>w ۵y ?6<+_k_9^SoNZ0I2mO':v 3_:kN l@xi[v[hI)Ղ~aЋx Z5'S4I|}w?,¯e~.IӀ$ 5'|,FE{ ־+Ķ?N_^1<-ӭ=<$OyѮOO8~Yٱ(<#G h`%՝pA"jd)0W% :B2~OB3h//=nx k?x]e>!DqJKaT Ӗ´0m)L[ Ӗ´0m)L[ Ӗ´0nYY\IamkGnsؑߨ)َ NW%A=fs[6e gs1> UカF׌A:yT$Ź廛z=識 r#4>g֛ޝ&h)[}cӇmy:۶/?KBh㡧6{>~R1CK=K"=X}/jxC=WVDTu @D @8e%~wh{Y-g̫((Ĝ$") 4r3@ `j_kM'!&ji.*`wItOZ#9ȬӮ} g@#_vu82 8sZoMX,uAQE KS.Aܨ :'t$;: hR pT46U=! s2#_ PIcSiINhܸȢ^H#bEe b$b0\8scc II@e$gSG Xs 1%o^oٌQBK!3[ GuBs _.NK;ȃd=G1!L7u?M+y䯦I4k-@ L}>`!W#T'v™:12b)#mf : k49z}F %&2` "Qg<ULhSuHҶ#ꦍ9_]gu9w *pS]VQ'IT8ɭ{W|O]3tlr팼Nq _9^S׭4zF i$Ǥu 3_i;'\8! ^-:ԣ%;ԛSByM5$ mbIR ݈p^7Dc8zpDt߲ u:6w?6B03g7bE e.Ar~PZ:01:Dt5JQ#E8(ƃjǹ2D&L M=;D{Ζyt~4PQcqɹQ~ ߵ>LwtIy1;edvFgSRGT)VRZD@d$=x O"(1<j)B|.G@mK)%.gJFjf,,$m@JrDQr9@ܼ%q-?{<"~̺ gb&:@+$ 2؜}=zR%0S G_T6VF1Y3h'bN%1D/ aQ |G #H)kf-Ax8@\8!pQDK+-Ղgȑ"m|_l&{`0|&BdI#əx_[O[-)VRpiȮ""_$ײto2!Tc/? oxZAד^V=]7"+f4 ՙD?6?3O^E{Arɻ7ɢOҽx޾_m'/f\_oÞB)0sf}ޢ/khլ~;@H[KRb{KmͰf kmU3opT>`żB@OWm/]3UFvd[}5hVy9fnXu4j{;)퍚vTN?l|ͻW/{*Ż~w/^yG~//߽yg?̢ϧ,"E( ?B[MiUޢiaӚ.{=C]vݛC YkGKpm@p0|ן|zqk>Uȇ˭=4zza'(6e?bTBM`4z@j" qǚi#C.tuE ד|:lp\)-&}a8'NFgv^s4 6m'a\|SI@w hDhȚS>pSKMV*t1vHWx;19z{$<1P-4yE l -7'o$.;o.Q}VS!;`AX+CU)(3ɘXw;ǵsT!z7yuEǜO (fNzcpe%)Ĩ5p:A PL"s<^v!<6Bɋ}"F)+tp.O P3 AR7n)JqBQo99yFr1i'KVD 3ё`|]C ԛahGbƲ:Ys|~D\ D3ɰl* !]+7+`G0V*#5 riƴL/ϣ!zobcic\, ox` Z@ؾTO2oq~ܡEM{g`X;IP& 1W5B;U9Ҧ*Qt6_!qU$N @c x⡯zdIBT*z(0JGH 1R$".&c#z1JNHޜAHpkvvfc|7;(ODŷ/1|h2mȜQ RVqFR, $*8h2vFdDmk# \F G4:B$"ѨEγ!MO5_M(~Q4Ine,9ATZѱڼAD,( zh˞X,3˭Lߣbg!hiF( >JOQI8oG ?*N̒dbIC(*r-aEXB)Jl%Rȉ5c;LXjq̘;K xC )e2D,%E|"e_m*hR4OOӧܴʇā&C #"$~kwpѕANG 1ot (P3g'?#A!(F|y!3"Hh@Nr(]4IqV#!of'oP&A 0%wB*$TF ̊hXQ"o&Lt\_*ח^_:*%7p_ioRRZ Չ fF!U=UMEnӏjEb^մJ^ +bTNqKKPP=%vP_SE,c7s £ϾC&Yfka6x:E DK VH /F7Ce]i/7< 9aMo#u_\z;WOk~Rr{Go4(u4z4 H ya0/ 0cY ya0/ ]0`^̋`-lRm, %ȯ0X ya0/HWB@a0/0`^ ya06JΡ_4?j*d "T*@(] JA1Is89.jNQԜ(j5JTHT|XLj,tJ^"91VX/%R1HL451s8@,J@mrpkcؙ8]6s#8O;-G~z?ܫO"G[^*D19Kۥt/,ޓҺΌ잪UƳV|`89TȝM vWkOQ )cv+,]?һNT,yqhb KVsm=]{UYbgԼVr>LF-X aKW`hXߍ^XF]lTeYu,?}EW5@d[쐽uϭ'͵lN0e|׀蠿|{v]+z b|zv6Eb*)'bbhSYm>AE??ymo>~NqsǟEg%fW/fxǵڢ6Qm9, Έ(Jvܔ,Jvud+[;+Jvd+Jvd+~RW|%;_Wd+Jvd+JvuPIIŽP#NrMM31Otp b_pL,`C.pZ#2:,2_MX%@ = 8il;I׸ ؅ ql z:ɳ&ɟq/Ad:bࢭ|m݌i/ |fI>[LpU|!XYЫTlpE-cd8m֥g@ IpEaٛhaػFбu<}%kқ|wLIKrqY-yV ZɳZ<9>9HɳZF7iIz.WOYЍLmL$ =4) JiWД|a'Jk˝i|<-9&ɨFWr %#īnU+Bg~-5`]^o[Grr}B6G36@~ n_4PV*\?`&vPMg$7R8˺k  {;uv@˚3j77n;& ? @Ko͵6g?*.Զ;lp hPv)<ðy*NwӚ[<,2VBԀg}@t}ges&txt$'=."Imrp9'\1\=c 8ꇘk i)"6K? B\tXbh@)q ΂Ҹ8Dp0"ބwj_ sz}c(!sb:%:?wGu`ۤIAjPKT2CMP EDU c-VihJ٥]pnVsc8|"!U.$q %Qd*C$%#" IƠMjqTwO 3Ę͈1pYJ! <$%R%gGG)*3:Gbh2 6 .q)g7+rwjaQRʁ="(QTmHIMR+Yn#I"s)}\2m}Tꯟ#.A,YH$<Y z dA D8ꛮOQDL-hg,HUƓ⤇Z%cNօM,i냂?tXs~sVIVU-wTM魕s!DRt(sRB D8}d #avnE);IM -ڒB##+V ^H/"IZZ)숹{=I'иXt):Q} ` ]&|J'{X妤1Aێ+t ]Y)dW! Q@SH ѥf#jK7'lEO9RX3bVeb {'`C_u`;<V(H$[@yU벫'gXe& BGW0 d qai YezzzK05eB Br50/GPѰ֞F(Zh@#_;wMA[S!LI2', )l:oA`KU ݠ * 6`!+J$W{HT*#``!&S`#2CE (  LĴjyU!}ڄLkPtq^X*h!y |Y%@ VZx o!aHp#+ѐZhFymC6YK1Z1PݩyaU0>MiWdi"&"MBU+WF0=uR=<IrC*KtȻK}C 0 >Bk)a*2B=C%KcZ AK^ Bkڝh Ō6!X[vCiXQ;-,:6,k*ZSR'ՓFjׄ `LvoFo9f(ICQfluAwP)HH ^Q%fDi OdPHz)q@Xolԣdؒ7v+TOn7;dE\Q"Ndvi'7l"g`" @+^F.LtBE9Ff,(JFb1;z+~(] ڈJ5xQ3r `NmS;Mkf E54&EH4/Q4d Z U-ECwsr~ecB]hADoZ ajlJȝ5VA-2m@ǡVp* [6"i/zB_ iF8r6G ʼnrhj"> 8%hޤQp `~E!⼰ކ`PM5r7IDt9\phҪ\ .BFCtw]H蠀%Z 4|/Hm,?Au Azgjp!G(+B-M-n>?[AJejQ ԧ&;_mƣ1(-)"BLuFOܮ~SMgvIzߝ j#5N qA\?L3 Ho)f@[obwVnC ӑX>D/S,OQg;-_n9ݼ;B}Gv.^\t= sGP2&."Qrd=#Ra'oV9ڍ{"7Ebs'<w ݹ߷Rt{.jjx?M(iK0{FK, .iV( ێ|Y\mj#ʫZ0q[^7e}هLn&1SlJ+pQi=&c{Lp 1=&c{Lp 1=&c{Lp 1=&c{Lp 1=&c{Lp 1=&c{L|^AڏcbSqzLx~AKQ=& r4{L|XS 1qr8q)~NNt@*rm L9=Po%H(I;JQbucqyrvۗ9Y|S9ON(^co/`PE!<_/k+̑&n^/-zoy_{XXx?.̯REO`]V˫7˸ciSډٻoK_D-.Wk;`n2wi7#o/ l u),\ Fi÷!cN/8P?gg',r5 fcllog2kƆFfcm. L3*5"P(=&ԓBhؼąOrPȦw=>v^=y{[P`MtH"-Mf&2W¬,.|>> .ܺqr}wy'R I݅,RY.XNu:@uő8ݑj5[5 -(<*ps'Xt+X# 碯]о8z8<=o|03A'F?]~StycFPg9CZ[^[S:llX;9;+ct5ЁxX&l]~x᠌YwVtuݯӍ)6i'pҮO5cط ޮ{=aӻUo;q/SKy kC7x  ݠMnan@nO*C7> ڍtqЍ r/& #ړ>` Mb a2F tLD*[ OGy4s@i<=3c ?b&Tu!Ę)v^y^Dg{"jסL&yDK{dkqfwZX<ۜgQ_}C^hP՗9go}?{@}482OmxV,'O;> O !I[tI"Ȳrq3:oX2t@.Jҩ"r2{GDd=L]v X~?;{:s>X3rJ"d O|8yDq"' m.:;$UO};y`<9<j?L6#K&3ٜL6g9͙lds&3ٜL6g9͙lds&3ٜL6g9͙lds&3ٜL6g9͙lds&3ٜL6g9͙lds&3ٜL6g9?ai=z~_R- ]AܥP^^3VJ XˏÍF}nj4 }(/FkV&ed"C VC= q8=a&pf<б/|U_x-=wjuHOu:|u<[/Z8[l?FfbhXQ5[ZJYFV%URdՍm= T9+mj 5,0իX=@!veb3u:/>n=7J;/xa\{m^WխdM &SM騵g)nFB)أz_jc37lililiii^(KFhY5?V[a{w~H2]|suDPu!=iں7i^WgO?M?xx FJΉ>ݸϥ}ݛe |h_/޹|qblHW`HV #V3ɭ^: WüJ^Lai*b882[՘M×*Qg\5ʹZ3q:nI _tMuo{p_wlMo^__~/7ۯ?Wo7/śW?}O/H$L?v"j_rMWCx%ۇFƺt7Wv9d؂ n ~-I(@/V\f]{IT (Wlb7K\Ne_ݞ%w}&tn.6σe"IF/\=kZtoWQm QᾹ'z3?~E 4}ߝ]5k_zV]q{r3/pΐʹ"tnxPYkh]_]"nm!R(N\-樌bϱRںokZ]GJxyIu^)\) V*Rf#CۡNnFiX:[% .mSёϯWFDݘtZW@Bi*2Gs.O\!W+$E jGXdN5`mqKϐK?mwAgʠdAQd4^ h΄ߒXk-TݸBV! ql޲8g]au<3Nhh2 MhL $ʼn/AI\rОEv$)ukuޡU9[u$/orÇ4fi*?)1ΓjEnW?w2H+@M %*o=u̺ijSStV@*y;aީxԾlZ+MnqWaQfJR-¬o;@,7yY6HiZbAmmEsRYMq6m(p$JxjQ -R6Y;`Xۤ_ug/:)XvIJNSp [TYQD<3DYB5Hkfڥfٵ<{]]X>Wb[hRg}Srd~3!F׻mZVwo޼zw`i⊩о,c去[|Z/x3n}g?O?+ 2̷…dܢMmQޡT=#U = uﺭ1i{z/u8Ϙ w9OFڢ;ppc왴D™ 7(>DY{sAܠ64yTi6Z?Ius"!lɯKK7vKzs/ Ҟ1^SkjxM 55^s25wPJ"|TQ>*G-^>jQ>"|TQ>*GE"|TQ>*GE"|TQ>*TIM>xם u7r>lݨNYvC^[z"|̜Q>*GE"|TQ>;Ɛ7K:|6#X0j.TR sF5 6+h&oZ0ȵǜkGʵr͵%-wRs$sLic|">!YuTœ$C iBBSФRҐ2m"Rqcpޓ`Q Ԣ5Y,xm뇝UD6w/mLʊ`A, ŗZWnhVsמoeї8}hKSYt1{KSxOPNmLZF>wͅ+KqtFdOE[} n_Z7뻧nt\_mԲtnƷw>٠煖pTWsdϓEYՍ䫷t]0ɴ[Wdak[s6YMV\osfn˯3h':J}*hU@PӚ1WGʹç+>見 St񜧦d524"[@L28 " ?:Ve+;l\ .sds{:̘1eEH%0*2&lSF] uQ;c^Y"g9ȀZNPiHJQȹ؞E_ ˑfx/)!s1z3Ƴ9@5FB`>{QX* }G)-H:-bAष])%ILy}vdb&ϢuN˅f-Pofꐘ+1Pw(LsL1"C&;He8ork:)4W ȣwRJ|" J{I{Iǎ.$7j~G0Oj5OZ ø9R6 tyP!7['N+$EdĸYd߂6jQƠBF-xnbLŠ VȊmC9_pA{-*öjZҴJ1%}ΌԧqaƑ]Zu>wXр+-(mxE25YVk +;-3LBNdL VdS,cV)%.*ko/ELu1:t\H .H Kfu X[D YxoGmJ[Ǘ.sz!+v͈vakFؽt2f`,?s է<Y!Hrn<(OH'"v۲cvНD;SeW ;9 a߷G,PSÌLZ2H+9 0&u!30(W :Ky"Usj.5ŃgOA)%, e,Q,jHF~<g&*)il01tge'B[i#'50LdR`"}P1)'ɜYglgLJw;fclAgK/P٤$b(T N6 WLPYKaOrq΄ٲ裓B1L|dRAڕd1Ц`ˁVgVW;I NH\4#y&$ O53$,0>Y:ju#8 z=@wVF5,k@e4撤C`N& %#S$M;/z#oP;V@J`zѣcaζȶ̧:f9 |i~L .c>8J4d9.(I{̕yţ@jl'xTNЏrw"5i&ήU$J, (6Ahv)AJy 0F D]Tbz}U͞]hO̗ 9ogQ61`YEZQoazyu.ѱIfHV"w\v7wթ_tS^ùyC տu$7q}~h»zBLq(A}dy@<0`^(!chi۩3V0=Po<$o YQqKu0tl+wF5o/J=t˛xd>*j{1vzw`ۓ*cթT=JIfj`kr.3NB] uυg]_A<śG66}|4۽&۟(1c#Kd0s+QN$ɄM`Ȃx\+Y hu݊$e(66!hFhh̺,%p:c ;i_q8;Lqù kWڦgނݷC[wBaJY d NK;r5i1dTìJ2@^4dkbbE5QBiZ"CrSgs[4 3+Xmʈ2#ڞ{FC8 5) -ke4 rA pH!Ho8 WƩm7}T#GZv(wD}ӗGu cǸmD} ٕ@ {ENS[GL=<> ehr}q 646r!e}DYG!T!T'!QڤQK6[VYb2V6L,#Y[:\Zz6:_Y꼷6;_=|k  Wzuz^S4FQ3}喷0eC)>!ӗfdȝHy'ǫ33qO>m}_z+QwLu/ʢ7=n:n K]^M&ƣ`ḋHJRh$pQ{5pS|mgvPѻsmFvѺirW{yZ4b4 %w침3֬g=.2{`0<;pAƠx:1^~vlcW ST!^RkpX@XBVҶ58`z96Rw޳\Xngp ]{qBI/(6_iEgoPeh_X|ĽG:}xv6]j".-^Xe*m-G֓јxx3XbEpQ#} )^wJʺQ:魯SʭthS|P򧲢fˁIwn^V20т1&TF 3 Lt1}A+>gf\x,q ɛO]TmozfS꺹BjЖPR~ 3dMl(M65KUȍ6L5!Y JFrޟӸ\ Sr g_Y ^ &iT2TzREe' xV0"x]bį,8[ԥlyHnǁ^17lR|gX*`݇V4>~լɟeLTutH2enle2EjiQu $q"DfJbaoU[-+ Yd%ـt!K+1tbکLl$vP )G|,+Vrΐ5BȌ,B0Wm̵Os:$n1tC8<,p֔6`/e{h[? n|&LL-7TւLi"1pO>BM .Z&iAiH]Oq"3;WEWEZiNWEJ՟ٽGVpwKtE ]UxOWctov""i ]"MOWthm؞ɲ)xao˿<-*~2Ys@`؀xft}46_>pB32tL:1K @r+AJ :h:,8XzYST>ƿN&EXsgG#ZU`c>?}_߆t~{z%KR+57wLo =-!ҊLȂ"+e9x4BEWd:dR`D$,T!.PLl@}ct:7F4sHd<݁ c/ONuP^Ԏf p2ƈ^8QB>o kL VRW4c{ћS=hWiЙL2lp9!H87t 1/[MÑC;LO;SȵHe")q[Z/gh :]yS削R\OtLkov.x!>`!>HNZ]ɒfR#hd٤̉ ]Q#r޹q8@7+G1I#Z%d@{LyJơXge/B;i#'n 6 쑼 8"%cI2deUHOUٹ=~,V2-9 C}vt%YώN%3˴C IIYWNE%H%V5S O5x] %Tf`*m0)**ēJ `Ȃ4Q6$ghuSH퉓gHeQ3,kee])zwLV'$,2&iy1| !cz'"$ijJkѳ9|P|Ggks@f&7ztќm)DeN@kG=x'߻j;Tt ]~Ŋ~qTljhHYe|֏Kwi RgSGyMfUώC7%l]@~nSOF5hh2D Z+JNW}shh;OYFQ>zq8|Zw> g8>L*~߭V"6/P{cYq!ߊӊ>vZ.niҼdF,NbK! /<ާ`N{$5~ 9uxUd,ڭ}vW^JoOmxULѥZ7ԆٵQjmogޤIedG${7ͧշ7g3`5O^hB,Ɓ9`,!Co5]xmӹeӡeӉ)%Зq.!ȳ` LfyXe0_ .rkNv_%KtAH <YkKQ\Ȱ8GteYrm126ga?PN5~}3S۫WWt*+W|ꯔV]'[Rn.#gr9JAu{ZfqOm?^k㣨 hr}ɤUhJ?A{}UG]e2j!MlEa%2+td{L\]QscRrnԨ 4H )KnCL%8 ] 2q&\e8ocC_߻}ᷥ?^v{{TK+{)/,uޛwMkyC#al3dTtA%쐣b ȣ^Ex`,CT*qR6x\;OMN!xbrD%B(OjU0_8N&+moVkO}+qɥYS\?WNHe?ՑrNet2_=▝ˇu@}8 O_rh7pE^Ŝ~. #[˩3Հyո.ar/7mݱs~qY&[:Lu7=n:n K1)A`< ȏh$V^J7+HGQ7E}fG z #*nLt/BF쟦aDn=5zc.[G z{+`{bڱ^5NQzJm6ai`YWYIข堗sJƼg.xt/v;eqBIy?w$E-4_iEgi1caf /,R}l>#Rl><; |t/,Ri'U ux{LFcq]XRfmbgE==_R/lxީk*m("D 뤷N+ҡyxq9&?@1+j?qx+T8}vyt,c-cBebHI8=cUJ&C>wNGU/FJ1 ʮ:(X;1 F&sۧ{4.;h*o2%CC/5>89x<֗^b>tܣD07I_$}-DUvHd ^tڧ1sns6am\`$5'8X+but 'm=#y]N/N./|9E@x,ʕeF54@LgB}˿cJ]Ԑe6XSP\}+\d3˜;mЭY9{ؾmi랆c]>"deLxͣ5xٻ1IC/+vb) Ӂ i/K4tU 1BVq cm_̶EO}m8(~m 'd 5K^/)Qc,-N׆4=5)k/q|_gv8B u)AA^}@ltk kςYc9-?7xr}޶Έ"~W9:^j>z|,wDl-{.fOʽpD("dR^~Ae.&Bփ *mpI 0{DN=Ieh|%uQo R>$xNMзo;G^kQʲT,k d.,B6YK b`4{mr&#Lcq@"d*)E\  I8' N$4u^ O:0 ) Q2Gza|+u JR Iz=>jv'<6 E{=ηtu/<̝t[</ht KZ_VG vSR? jkDrY-fϳq3FJT¦3Lc5ˑ?R%h~z' 8*r$*Gcz.)b}U4'@ǟ[u-?Ϫ¾/gzxtv4jy=!zG>9hrr?xĖ/}z2}Ijܼ_ѯ1O~mN>x5_!F1ƹ3bN/w;'2>C5}vQDթ&e#it$7tl|'0oqT%ާ,Xv~m:1;gѪsf騂uF]D{6개^ kOA,]}]sUNZ~SVt'ޏ;Ə9?%v_~|oo_~囷_Ƿ$m#)$;;6CCKmk'ɸW^2U)P.םa,TߦͭB̘tf^u/Ui"lR3a7SvNц2H%T7LT \}>6:gT% W\府q'C^;:[l2KjXX3g/;iJأ.i,F 1y0Z[LJyQTB erkY#suڅST.M稳蔫2-wSV岘bQe~=!aszTUߦInt}Q沥ާ^Koəmk`^]"gyy{w'>uQK,-G Xz|*kd;=sVYs?1%˔rPα.ٵ EGU;wܡ򏲥?\tRV VJfȚJy#h2Edin`Z>ۋ,gQ"|8i\PY&h-i:JHD&[>k ?O`y6f`SYY`>I|}ڋAvIRcH-nhsYCdiISȂ CpXrky5'Ζ7 ԫ^$:"`lf{e SplXĬ&{3$n|pЂKCR\jlߍhV:p<)a^JiT@I;%n)-1AvH89 G'\T|C?̡àqJRQB3{Sp\2/J ON{o*Oz#tp? %OpTMNqq!VeVNΧ\o@SJs圲HoJn*X2 GBqϳ h\& pP`M7%lv;32._l#8 iMԎ;awi^pPnN[Xɀ/ҵV֥W|5wБ. *:,1LJN%4*z XeXDTYBɉGc[ )GTl4R|&;µKNCNВ=jU2ZZs+ymL[ "dlQ"S0m8(|lD! l! )H("\}I hnFΆtگPz[ 6.y$r%i~I} Zu'Ctj]SKzzu4Ngm֔ճtH-kj9-yZϗw>ߠ繖aotG;o^QZw5'gB>//rrW8oXsӦ:-2b4V=lTm,m8bR[5͋l/˕$g}Dܒy>>`!+WQpn0Y1U?u~y5K3~Iz_&,vPi"ϯ %5!Lr2ZZ_~cM^7tR`xW&q\VU"4NHOUT1ox <8M^s: *Ώ3m%}G >16|x_LX`GȦMv8:/&"p[]%!l3%t_8{if9M4B$pNƏtk &!v VcH©]x='ݮu'o|q{F|k2m5 N0 ڵ%ӳ\-[΋l{ coAMF!W}(ZBs'LY|n4 c0R̼.y$j煪1Eв`]k;| ʁQ9:Ӻ!Q|h6x5Zrvܡ&j/ÜmYs ==m m2m2+I׀P)B&-p;R@_z7Ma Vjsʧi!tl`6S/GsxYlϫ4|&Tȫu/}Yh%Uu?{WG]#`yX`gO[۲$dk܋:TRKe僻,f_q,ĬOmFcM~;=\m1+C`UZΡ.)oSMηf>{c,۫_sh5/w܁ѯ4wӐnXn;Vń߄= =8~hѼətc.3dעs l`4KUJ5ʷdfqf[>糼AQpWqN{/^Rl99 ESUgSˢJqMXh8UPb<*vWUZ;4CW? \^$$X@?f[.ʿL?ߗY(E{ +k;/vщOH],%MY/ aR^gE,хPJ hTTI H`^/!@b.`ozss~֧b>-tmL:*w0W}a|oSZ'֒2`N"ɷ[ˤ4[ˤā]l}rzMhW[S eR$ .]2w(uө Qk>aiN{QvHiFO?gS"4IFZ fJ9ʢ1hPLL[IAM3|GQ\ Et-|Fɘ()g"ad "d ٞ{^2OQ9V]`&`C&r(lh2"7R2 VJTW QRQ%hҀV`&O6٦ 4%9\u8;f_Q}@޸ 9yH%,chNv/6[.r!-~Fd~jK4~>Qlq0e=lDN,Eo18SX&gĔgp-[G~N~'Rc},mO>%/X^Mç饋6F*_D1W@0Q(tD BBҥ8Pxm6vʼݡvꆊOG.Bڞ(?KhF2ӟ՛H}ND}j|ќyɖFlWah ?VnZ~|x{Z^zrz׷7vcwY7e7σ(O4||sM#Hl=oҜw^gϼTTM:,8Z{}J_F4(u{T>5B |TQ:Ruh:yʉ"#lV5 U^hAZI()TbZl)"cFb6 7]3Ezao%0` sL״\%XZg2'S Uǐ F 5ˁuJ+qCkrYX5\p-9 oruCuS|BB*㢴>訝&(Q"h|kCS05B/Nsd#S9$-u/-sTv c ?!hGv#iǛ:2kWd$Du8;$&'r65JMt:\UZCϕ^v̕sK B #->w.U#Ot-ǭL71RLH[3!=[gcgA,њtOdB oNzcx9s?7wד?EN$1t  { H9p9pp8 DN\ޣE$yۃ( Z@'&)L=y 3"b`WrIEɼ/I]-2(mw!Bd=x9Y]p4g `O7Ϸs ԟgK_o>O54nt.&x#5bS4VQ-bqS3$f'v>3 |m-4zJfS1p{z+#aF`TwM_WB; b]tJ8d`QfK(czJMĆ]0) ZCǻY,gel:*@YUSkS8m,&fQ$#q-,9I5[@iVMJMRR1X0T9Q)1'+RA526g;2nf)'f+c_,ЎXxT,|RK= \kT&&qvzt'_`trHh&D-0(2)i"jlۛbIx:UMm1Z*IsH+q#vNKPP1j#ݴC_wĤ!J :0Nx49S LS<,ƢYP[QgdIYDh8,|H$٨['j&v<__}#"8"FߙtAvVƢcT/pAkg T ' @ޛ$͂XT6 E*gB gf$S P{ۓZ>e8TVd_\tqэ8O C׺Kt!HRIR%$&ˆšaձ'*ga 57CRIѵw߬cXg?ÃjńV5;#vkgz0E8c4@q (h5*&e$eP,- N(!gGm(P_`AS` %kubt-R9jMϳadv" .,Xz/.܋G!ߛWNQ^hZBU3?^͗{> k$b{Čڵ9Tbbә//+tS覗tC$.sDM)h`Ġ 5A)gRXTrT@ dZϺ, $NKzb. a0&=[MZ8b4!§UM?gZff ].kP,Ye](Ξ Q<1QC&n`jAcWb'C@Iv6aq`>;ّenݞTn^\ xt \q$e ͿrBu':{BH{fXx 4c{D8&gx;]"( P3Aj2l-Ԍ:Vp63XM"E=Ǒg8n3t*m DIѾ}5'iVn{Pq|WZԬ8<v87uPAybd[KDcT䂵9)12ؔl=T<38&|__Egq̻Uzϫ7WHa)ۡ%32:_|8=Z$<4D[qAb"&Fr13sP Ds ={@t>v]DPx;ŞT'|!p2wd5EE e.A.DrPZ:0Bݖ&y.4궊)A6V;E!2a y).*8G7Rϐ }Wz=-ݾy|ZY_i;>L[v05 ՔG4::jJ$ J)R"f$=Pgq'8U\|ʏgk v_QD3xddO)*)MB7ZQ"EBL a#ݕkB=H,r8L 7jM,u, DO2s|y J /ClAgAٷ/TBJI+|F1YP3TD< $!,JAri!n >nN)fqŷh^#".9"Q(JqK V$gɄ8v0_j٧mpek0f[?mpg˫͏*f4 9?j.?o|(}e4.V`Ft{L?Etv0P垚у_k7?O~_47a Ku8Nwaǫ mEՓ+ؽe2}Zr88Ը'O1Kxmyг|j].7ٶj^ jYvZǏ j_'SFz0Ii=`jgۓjgR8bxxW?~?߿ϿǷ?H}|[fG U+ An-~Ҫq\5!]9c]neݛC `ĵ4+\?_};? g>߯[ޤշ 2cC=fa8 +6墼p6z@Y W^ZF.44h2l#O)2l ?M5[nVy 9כnvon~ E7U &ܟ`˷!bl9@l x͍9VsԈփ-]௷^5P >]߾Y39ʐqos.i5n1HqKknmy0vU&h]hJ <>Ϟcbm=b)ચ|~Z玳ARz8ؠwN9I1 HAs0+C*03W`20k5ߝCz^4yu5uc9j^U' ŦB|)g;?>Ou4T*Ǚ$%L utEǍ{pgyκ/CeZqy#DW, 1|O9ґGIi+qı#ON>Ԕe9sփ/:|>qb>Bҧ2rFcS@ yu$<1kqUmvMjItyj/|hgvjYG-37n6ھAMFqi76޲ y:dDU-pJL̩C9ǜҝZ> Kyn*(A +54rg4D P)1Z\UE\򣮲Sn_.`ݲ(\A,ƅ_.x0qYd~Abb}5l}ݛ̶Wx7?'_-> Vۥ)pj[ ^]&=A؎qf~^m=89xc펐E{Us%ᔺ!ju2TJ ccûe(7͓t Ig>]!ɰ+Wu**C *C)Ȯra_MM—nCKN .uE@nGd&F*q2]0fhJJKh%(`UȓaWĩ =P/]ICo 7oY7 *WMX{Wy{kי^T6mYVT  9R#FRٽи4h\~2{V M@6:<-OKR<-OKRU<-JR)xZ i)xZ i)xZ i)xZ i)xZ i)xZ '%bnP)xZ i)xZ i)xZ }cb<-N)xZ i)xZ ߡ&rzU3Yeu4V922"%#[.)zKۡ;`y٭\DAs)jkΉBk*DO&3X0I5/TǾ=O (fNzcpe% uF R hǾ9[F ]MFCs/= rri i(!4I'bBhA?:N%֧L*H(pJQ"Ex0z ɑcJN$<f#X+F}#Awh7.> :Y+s|~D\ UlM`TYgD!O 骯/^+`G0V`J[j9 \?i * 6M"wh~@ka~q;U@ngPq>;PS*RE .SHUJSĊZhh*ah(~ _d"q^} uqqePvX#+cCr4M3USmDZ}#T:)\A0[pAb"&Θ %M9[w 1@TS&ktI5x=ˋZ Z-~ռm<@m6bvR]٬:I^!ק㵫8j1zxفIWcYO#+Y5z'onďyc~|N&{[>Ѵ18?׺8xhY2'7v'xk/$R)d1Oi:>wl+^oC7-b?m4H!}(1Uk[bU\;fe%1u-H+$䤎GvØb p):ٮ+pљ)AveGuN+A=QwJL:i%'0Jkk.=XQ$PRX,VLsLwG0C$ P*|)q2DJTWƫ;3Kq& ȑj[EXB(JL% Qi)]1uΑtֿ)a= *>K xC )e2D,%E|.NL8x(ڨ'T8V tI0#s[C@!˃]Ko[=+AV@ UbA^ 0HNOQv,ɖ)KJH}{:kvQ5&4搜b>f9&.@p*tvppLSiݹw^ϭ4}TucMZ#5bsBgFUlX~Dq,$fGr= eWo\)v'?.Loќ/.zu`8fLlzY)+ ճ39htusqqa[$쿾-_?ZUk n717@aL^Ak<L0frCJZкBj0o `J`[4g{P&}hTz\] jZKEӇwBiޭ"g՗CEv6^Ͽa7nNn4A. b_:~HkA1%k#V5EPr!v6a  Df!) KPus.%#vnQ{[~L2۰-ƿe>b0a0xB7ջ#0W6cXӆ%ݫǶӶ]z֖ڛ6_\O?}vlz5$@SB!t?q8eSB>xN'Ll}=ޡ%t 42Oia]=Z|O3[V|LV3x=^;}ҝͺM;k_zr=^^۾1( oIo`-b*(ĜKB2)R!(tKs9CˡзyZbB͟]bT+q`_tXu,{ܞeShY?^T扩ߕ&ŇZrD|/n{Y Xp 0ڌ麵W>obY`IT\BACafeXIC<.A4ƇKJ*l}z<'$<S*=t]?_9sA/Qe}mV6_?nlewWg86vdǎEMEFIE:ӛ&Zjr;WRQd>q}-RuZOU4EnJ^LRnh7aM -x˪zk ;5*dıdA'g§Tf>Z@ه*t6\.,-*zebu1G"*}#_x&w3]1꿾ま`NWw/#_ϗ3MJfgScgSIMP@MD1?SknFoZ \{t< ͪȥHRmdc1%X'Ƭ*>/'"}-6bHև"{.blWЮBuJ:j/`&F)[{GɻMg%=CSr^ Gܨ-|jo!# Cڲ'U;洽s,IgM>$ t\I͖8ϛӒd5ϐd5Cu^XXg^$?%(5I6dx ӂjo:gr<%)j¶a58ƜB^RѦ=Umk֡m]kFMg?dc{R D5E(8JDEhԹ+b} F)]-V}]N:0-3h1G"ũwA5݈oTP3j@`(KZ<Ҡ7AdljxXQHQX,PxE%g8lBJT 3vn<\8rǾȝ"Dܚꗂq*%jTmkΚ-XbC2//&%"o\569W)  HC2On3k:"u\uN)m싋3.xvA!S6L|(.66 jI%9 UqxEaձ'W{6v- {hYt2Eb}mː| JpɃP>\OWӣxO SKyТ#/r-59*sqY\9F qYRюfAXҕ9 ߑT:]֠ohN ucO辨JS|~A)x5{Eߗ,k%CaYY'TJIgCʜu72y->]Z;łxTp/bj %ƠN攩ާ$ݦۂYnnn}m>W X%v4sV(#;!3|w !3sp  %:N1a{ͥ?|.qWy0y#O@Ǣ+CZbHK i!-1ҭĐCZbHK i!-1%Đ vHK i!-1%"i!-1%!-1%Đ8ui3h# rߌ!<1'CxbO !<1<˽=KVSC6atgӐM4o,OCzcˀwk<]0Bi2Y)+{a/<׳{_kKj2*@OZ3(J'Cd JJ1U":J0 M;DoQR`қ}\] /sYn?}\ew]W_n=$_/BYƨIZRh[kMZI0AWx#gcWJPclQTk!hƲ%BW`պOQ^HD֠5F.jdM C-m:GGkCl(YW[\H}zVn.(ȁGxecQ!5Yi Oi|=/z 89}H(Tt$`"BB08Eʫj́ f?"7}jf?zFvMYgwwOPIvK=@jϗMPnu"P25ZqنZ-*zex:I0,IO)|a *Gksī. ;<}Bݖ3SXmwha2zN `kGjrXWox@Q1iƻbK/ /C ߶њppqgdiܗ_z+9AݒMԞjM*:8Hb}љ f cIYrYqЇ'O`z{N(~9B;9m!K9dRo.5b.ZrASSR6νF_bxXueLԪSEG*JeCsrz v(Z;:<(b0E6둲'yY>o"%Y`IT\BACi |Jt.AW|H^l2&ߗCZ'%lE頫w)zlsTe«r1!r0t2Nω$74=j_9~AZsC9,Xqdeh~ks8(2LcG>wv2crT*|X9!G%)Z&!x̃ rt\r(2hG * C0X}ujc6 *WЗ`EzR1[|y~ ڜfn>u>@1dZ[l*jofYZ4l5gȑ"m|T ef/ 63XX,y-{a+vKhIv`Y)f__abB!(=fћhh9]/rgl7p.KV voȾF(yg i|6IKf|uP9S œ50^%l* ,T >[$2J6nCԝ$Kw_PsI4&V0ep'mHs&Le4`Y͞i96vR_Q]}$m˓X>& ї4Re~=!]r4mz^u:<}ߋj? >-Kőcȶn˒ߌ j?jnߓq)@1n V)L-o1ln~zCQ,BKB}`MRrYpBt% %\;;S|dNI"6:S\iS!#v#F[t湕;)~4O謈2Q'|E Qކ C D)OԚ8-QAzcxg Ĥbh5GmyŮӚv&ΎyޛMEAd*Zi糳|=g{T_(3hy-=Igŋ%O ^^+=wJ#!# YBƨeʈAj|mU r]A3xU&,U B$H IVH A6kqI##EHJb jO}>&4MM>rAYwF9̭@*$+V&}P \H^TL4 ])g[ʤxvJd|r2sZDX)(DRc()H*mn;) F8/GGoA0/ 2-$EB|AkA0+mn$w,?őMӗs:z/Fe^pJX,(M!~v*EM./G>VYaG?>:I<5萖tb.?_gd٧EQS_џ1ߚ'oN]#^29ȕ0Gx'O}.Ѳ/kli}MW;IX[KʴxsKGm͈f376E3(鬞4e1NΗ=hs|4>9[[lsNWKju^Κ, #MNB,IG0^a*U4'~m*Oǣ0>>_o{}ǿ˻o Z@#pB񲴊o{ `\k M뮚MKۛ6uŀҮ-^UB\Jڂkdzo7FFxCbY q:Jw'%!%,k/…XF|(]_w\)!gXv}q~к}胲K ꕐݤz6=fW:=&X ~|Tuu_-\Ǟ.XQJ9=;Λ50zd2-{#*{ՌZl5\S/uW*{M%<܄qڢ`27FWN%8+HX^qEK~B.osưe"RjJwlXy-ֱ]v^p5<Nf -nO[;4qq|uKѥ:1zCHjj.,kJ{]uoz{ٝ :29SmtY`TuIZVj91%c@"H) ޱݙ8ۍel2=ChBi3Ƴ9`2!3C`}80HHC>ix暢EIED5eX%iT&0oq,J9ݧ) 8WiҚ4>k@hYEۘYU񝫒xTiGA~=4wvH#MiZ%Э8s29F{)gv@Z@ :GZW5B' ?I2 rЄ*fuaJ{Uw{{C=!^DG\]ih )Q rL`'6&B[ 59a1M*s/$Fg}}`F9pʢ`)*;vBu&źܧh}C[HkBxY\,AMm8= fW="z4t+E\x4t+EZ}[)RjV!݊4`؋2;)vuyh\E8$9֢2J"XetoBhr=BK\Q].khBTT'!*"MEx)&[M"֜ɆI#Cr^iq΢ԌIx́I*!sA{)BAs;܊PI-=9 @XGu< V($Qę@8CzXvR )e&r f"$mQ 2`FR>IX:gG=ڏGzD} FlAeK/lRq|1ECpӫMуq:GTKO_1 C_p:,gˢNFI2~gCIIЕŨd1Ф`k" P3N'4Ix`T*媞4A1ZJAh`'L4-sh4>YUȃѫT*͉'J#r{^`XVhdQG2UQUhMduR4>|>Aۺ;!J4P|M*t|bՃPGXDЅ!XI朌J.sz\џzt+0EfmH;!ټlM/HUY^>d5(rˀ!6ab〡n. ӓI 0F{c Wd+[܉eĶlz H%@x^⁗x%x^⁝n%x^⁗xLx^⁗x%tm%xq =nՁx%/K<H"gΐH=6Abd@A `%b~{E/bւoMC{ZY<%\IrKjgd@S%d5s%cttBJtAReD_l2A=Ƙ~V,:vt&PyvS]jN3x yfNjN} Gjߓ{hMp>}4w}&OUQu*. AŜ PrUZm6 021`31@GQs؋L V*GTSZTu'!ZC4+l *AIaoM)z༧hu2GwV:^`gl'h=?9o%r.pD˷vsP|~"VOi]gAmvKՋ]xgǧuWKNBnm^R`~ Rg3IbOCt*]QtՔ\= tU\7GJVTZݫ4j(5<Nf -n.s~6ͻE\7_Rqth6;zC4jj.,kRr4Vǿ yhCDC-6(\qQXhq},gk{~H~A9?{WF0_vwH\fr0|RV+$[v, !3RSM6Sd3L+?S[V~Ung3MZV~igZV~igϴ3L+?4mkgZV~igZV~DimL+?2!3L+?(C($3~dMIf:fM\ 5)'xN38e3 e3Ľ26eF1ș*e3Nz!wNRw\8&3B謏@EjN[%XZ[Նs ~ӏk $B.$P`i$ɉ+sv Trr&LۺgjZehXv/COJbT*zcgsH֣YEV4Wʸ !+2+yx<됪ԫBsz>OcUT9JeyV!1"+Y GŕJ)ɴ)p%YDtRTEF&ǘg-䂠'4kb^D" Τs.HFjùqjXXM22ꆅ{KgoIɵxzN/ù~ :wr2|2>;Jf@2IeERgy (V3O|uOfFBP= ) lJm!a!$,&:H's3b$nCAjq_6Q4n v8P! 4I#WNM@d*f XL̐YA&CX Z$Aq,59,Qpnި~JٸLCAjq_Dʈ " Nh@yIy2xٓ#G RZ Z"l%"S6N 4p \pGOztN@*(Hg8Ad Zù?'z՞pq.<}侸h+mpqģvLIu>D₣"J"A00!3t.դx(X]<\x; ے|N5V  MU לY[3c1cHptϢF dɝs2!SFP&dlF堭ΈI-!c PT1:2]H\mXDzObݢ54H )K>RX Bg.Dr(t 2yu r]>ΦR ?ƟV9S:P!Ν-mVkޡIdv@@tlo2ZT=A1#fV31rU]`9-ru% doY#X {v>C65Z l:]m0ί=Yj mݮh#Y(ӤF訂e2c\q͎NH5d;΅:&>ZJ$[FӘ=r JE`2;A$%N!Y;_~kr1;B.CaAӐ==q2[W $vmƽmN=bYڦOh1Yڅ ܉VCI4N/J9pD;FgI{A"+1nЀ>HYr$%B&hA\^s1($JW,u5Y26#"唂w U-6[ݲa oD~~7j)Nn=xPݷ)}Dt[u&*1寐vP'ί?|4#~{皕6*P|3 XrpZv,ǘ2"ґ!L1^+n.g8^yZ}JBjԓ元Ǫ!{^W uO z.e%]k(ړf/E>`uRԁJsY.Ni{h %t1FsUSХI͟`ш{֖A'JI9+L"L P̋ joCT )!&]iBb8{`mĆHtLBYC-sV-0|0$(qxQF2-}9YH%.ߜ*o{l+;M\;.yNxy9i1G@FbF:YBV(Q*\%7dټ%ti2"AB@J/  rpт  4"9R*`@l"bpw^Hp$- chz+ ane9"Y_2ФKz … @AѸ&"/ФPKm)g4-ƔN˒JXntfH@` K1mHZJc"gTy?:z{GqCLHH8y0F:9FEKBoY#70*r],UϨo~ ڂKpVϬVf_gAǥ]6wMVGe>@./:[x_x~0$.Y66 Fnr>'i֓ƭ=m݈X/y&^i,b:9EbYWv~mn}VK}li+,+Ngܚ4Ns="qSM2ƓQ<&qo{?oo޿{.|WfX@l̃`U;tmjuM7w|Մߦ_Y{U}8c@Jp $gPN߿SrqG{4.yIOhؤ槽VN׏JтV&/d.iӰC.t}?& '/7NT/]s˻y'B<;Y+||D/?1\0}Oִ|oFu9 Y6j'/II_r mBi ʉ]}sg ZC&ʑ*; 7v_G%[[ "$tRR.FLāy9Zjo1@1@m^^}ݷQ3 Z$\qU CBCTD YȐtРܶĽܶHwDMya{Ώ:B]yu jPax1Mo(eU_2!^^o a)+u^ i΄KVŖ[ge}:zP=?Y8CÈDW"X51}S|9SŻKgh$oU gO]r݂;kspU.uЮs#:@ilEs) ):f]$ڈ bb2 gJLȖtX뷠|4ۖ&V;Zۛs;mɮC̅]ޗ|15,7><;z!(?7oYW?˃̕,=l޵kE/m<|⢻ (+ɛ{8zDr<-Sx kDx~b.*> Ou4xƃ9+Gړ׃i[=qzwxqqVdS"~ [~ uW^PYק߯Y!fl]a}tV鍃ʅ!aۤX{f2TA6nQ{Z 0s'/'"ɳ".7WEZuEJ{jJ<ՋLKs|5`>^ҥP7*j;ѝ^=A  |'@bqu2s鿬;;_}#S˔O\3aL%3y۱ EA#o#mh !B;sc1 "O5^D#Yw Cv^6d ΃N*H0fYf I(lb 4r_FL1`FE+=蘸W&~o(AW*ZGv듿}~ߧ0kg/ 4?s_ݦ{Uxu/N%0 aR8q ܳ5!Rgdfz14i˒w؇~V'S:9? Y[xDxv" bPsm0k2"WXxZ 0 ؔ"qZcp[@L@pVg9YJ{*[y`睐}Bf(Óֵ?Ye؏O^ If:p4^r`tdEbz$o㌕3QCY'2m9CL"u $4I4RCLf&>%.d?MKSh]MOhm<=FG'_Oo%Z^n[;woNOf+H9Z>1Tޟ i,Cbdd)`$Hi)MSDy 2tz"2:ǘg%VR8AO &cGNQ@R8R iơX*cXG$a(dE΄\BЫR!FMF`f6U9L$>%~qܠմP֕Q[[4ޓJ#u[.CR8''N;t4kYFY*'+ !eH,: kbbAMPd8ĜYFT'\e<&~.}+lCԹqθ A7&hteowo Z栬$l H#dLp(͔"ѱS&U9rO(& (3,@2pC Ld*2j< r]hOz#z:s)ė>L XFh;C_-lT|Nra4}Ff1Hsrj2Zi1A28d'&d@˾Ϊd\H|@6%BLNsILlJ5q{`O'y|lW;P) W򅘥1:ܟ9<2* suý loh|NTbzHy<W4NF4=ng'?GRFd22wORB@CVKZKWܱ#FAGz| .Jdt跾88^6>^Wv3Ö$vP,+~\N/ә|>\ { CNGjzX+L2 y#YG B lS UgɑB+O&"1dd!%!gLE.J:tU4fNE.nF49~9;ڃ!P;U5qGJO;^%syȒWhrYnQ`}2ۻ%Ͳ5y6+OW<2fJ!140d1(S1nlRYyꄀ'ZxpJlBi#9sևe"rkEr_2I ]0‰wuRe7 *qB@k :k02X4͈h"$%dU(R4N kOa>ȿtx$W*Ut%KV ( >(.$i *F"?^۟kizDhz֠2);N[*3HMK1eOJB %G;D:F`‘y2Hs2ĥ 3#Iv>CZ809Fչ3.Omx7?Kqݰxn׮UjP0Tfo\?\_l~pkEe`A_-*`i Rj]68~Е3-[n;noW~\_3b-i8Σռ}z x2x=4~uLgO(CW7;|x>`ŲBLƣ@>90z8;{lwY-Mur-Ut,I}m2 į<(CM qSE:f H_{?S~O?{WFB_#QՇy7AvK O-akϧzxIHGl3S}ԯxυy޼?Jh/47)o$ALDݫy^z5ojiWƚ|yl{}8c+@\Jð5InO_:oP-Cr#a^zqMlqv0¼TY0(#Z6|23(9]ƷTUs67߹0-fǿ>rw4pOV%Mr9juAûV,_ ~M9 +!7 ?'_^kc9wRFt{?ɶpcʉt93!Ʊnl{?ʃ-?_}5/`^92~PM ÜJ~Z^%*%Go,PdzCCG1یA > K8FuP\Rњ$Z)TYb"Ιt,˸Bi2ǍڮewvQQ)?6((ӠӋ_H WD 1[ݧikgoen>0|oس-;9ݸv;4oqÛ;:n ;V\gLzJ8Eӻ=/2f2,ėKze|Töm!{UH. &?7x| s]YXlј$x(_hy/o] ֦EIDJts2{iT0R2MTLgl 9sP0 ώ8 ^rJm=X"gtFc@"I) ^c<[KcNݿ Jpiq9c<+&2Z;4 ${^Os߫ mq✢EI"f9{qVIBLeca <9%Wb]$T!1MvWzdFg)k-6TWmm=yiDS4 C2԰8|1ܢƀΤ$̣yN{{NN؝^;=yXh[KAvq%c ؓơt: ɢvnъ&vWmf1_";WŽO{_i݅|G 義**''cGcM;OQ1e QѢ:&$}9ݰ^os៣uaW'miKTrDqgrrUgrj-o{wysK 0\] n@VnD0+,¨ѕ`J#"d΋e׫<"<" ttg t*N1i\ŒV<&KkKNl]IDBJtA "ˈ \ez?cBkZqc<~4gͮ]zb>ݼ]w R;F棡>~W=EՀTi,Q娅IfkS<ɳ9}:'_l/6{G-DT721f>jsP zai4GF5@KIaoM)z'`Ef1Z̑읕΢W հ9[zZ̰(ꔗ<Ǯ 7+;]چ;׷hpMXbu,\>\ ont~mk"$m: RvR]=^r69]]~yFIl#gK>k_uX͆*]nK ,`LE'*[@IyIP3 NS tس@,M_@N;->:-&3˴ G&%)]Y*J- Bj5jQo'v3Rdd"(FxT:$ B #OYUȽ plYck⅔-۬jۈie@J: NV^GR߼ϡvDR["24|F:z:.GXI朌J.s~cͱR[fS+d_'⎓4-5f71oYR"QN qMHmUv(dlugB\-DWB h_4?}Zcr XtMP\_r~t6jQd  pGnƱ[v7[Mvoھt:m(-#Y%o>@xznzDNgsW9h_ BŬ@ցzrЭ-Э J{pBMIAY+  gXv L_l0Nȍ ^9}cdU^H0rERT27|9pqMtSxPVsϷ`ϊn ^zΨP>*X]fǮ*1z}I,Xӡ "X)ٙBXX*O,NBq:WD*jZWJ#z:E̬$]Q W%w|)WQ3|5xM֚䬇.' uWLgoo~\OnfFMRpJWιQJq %.t!tZd0]}Si#.U)k*Z"jk^rմӁ++Ƞ%v WZmBW9%2~tڍ`n9xJLqϳ,KX|VhZ'٭ӕ̳/c\B}TNu%l*o<JLseή'Q`:k v\&ZPo̾W,jlTXɕ*#s(@ 1^k@18k~XT9`_& hSs}\/G_m7n?6Ðe6C޺G_< 2nW E9u 'ӒT+Pq;UJv#Y(%;QkvZɍ;(%+%{ORJ^,6PDkU46\T4jysi+Q|\8nMߖ-*!*Ԇ1{0t  fJw.R-3c*8W+oZyk yTT&2D:+ո(݅ZY?XPˆ͛zsdY>ЖdI+sfLCfrC.[}8"T{qB[t#Z:YCREI+ӯbLG/HmQ>s,gZɑ_{cG7|Zz2DBb4E*H5=ߝ(I) CR(T!×@>SZ $ 2I.)*#|-ti/0HasT.-'TFO4`T|D!M]fT*߽B7@NCtY)G!(,1 ,f*lY(?i`Zͨ4Iw{:dӁ+$=kZx>ɠrޠ7Y$C D'eǒV7ׯ+Z\Mfqdb{v}oCg1^ގ6O}1Pf,lUVr(r9YYtLR*8j[͂Ǘ'zy*LDU>h|Q=C}@.5'F4?lE_".-x)҂JL1jDH(N@'L!d+|],`tTL> R!R7DSQlyf{@OtVWj%Vܮ| zJ%8}Mzn9<8V` 6CmdC;0?m˃%7%VɌ$FHq߾_?^O:nų-U}kr~rﯓeLwhiwDNѮD Ф'ў5(攪G<ًrz n XSl5G/.u (DpK );,8:,rXv_LD0R98^eqRXeѣpGb¶D=.I}Y (=[dRR&'F6%([]o6- .|\<:w fJ{U'nIWR6V dVjq$*5THY# B&U*MsgrSƘR$\d :l'n^ g7{ѳ}BV_Nd`}8$5nD$SتVwpm͐,emS  M@DIZրмFOo +c#O{&+jrQvt qSA)YS{ʌeO UY&W}2Auyjϗ[~' AQy^&x/ALVѨZ^ 2e0bR_%DGUSdޕKJҶ碥b!`.T})e+RA526ndlUaa38 mc,# FTmuS@bXv{..q5]u.N7cQ],4LPdб*W](dVUӤl"<~ ؋Y Ҫ¦-$ ٩&-|( A#vF4{&橠v38]cv#j 9O}2)"e0+cJtbgрgD4-9xLS<,ƺYP[Ѩ1dh)#c!\LF5:pvaoϿb1cɩ b38}cD#"lFeE)ɢJ,%{5h >H(}`[V`cŦ"mL4,IRq&zF,򈬂 NYjZ g7"^Oub\\v:XVRr(.ƸF\q96 ]VQUf!RrbI*TY3 aca38h^G,_+Ƚ1);k7q'@AxƘSיRGk8 uJA~ Qٺ/5w?|q92xͺKU ̮t1:"a])YSԭkږpr1bIxmg `dj(xTh伲Y(x6uEID͙@%S>eP\Рl*XAQ"ۜxp^/5}=CsxV/64ί/A|v點zCśf{kbv5]FʙuQ|pL&tQB(H(/dkd",S6mݹ1H^}R҆e$ S0!0nCLPa؉0G͆[Y߁/UJww׏kib!6lWaD6,)Z;rʣڑ ,qSz21 vI1*2-!2TK 5z$CJ:h7nS"lc:I~.KնIoo'{_/O6(5Ҽn#`PRQ! XH3tMͤ!`ȔRe;4 nD BS+:Ï >J dQYmbJ&aiA"d%$(!lVkgMmuw5Ϡl4 9 +V|_zKeԪDUc`h6=~GA@t*_>/KfxGAEzwS}xD|}xٗn΁ǐ4H5 2v) FTD6aW)'5O/Vy$3"<:Ɉ**kBփ7Yl0^im%r*Kzev#4AJ#N_<*K1B7q>Pޥ@+p9 dLl*F `&KM$@QJQi(Cj1ġ]z)jmMr_0F|W4z>_} M>*Iu,$"I&Zс'_ 1:5gV/gy>smx^%jZb$x'U':[PRN׳Yͯݥ?}qz}F+XYY̯ΖϾuߨsH@>6hqkG f}}2M Ԧt5O {xx>Aw:{GRܽۧ{!`y=>p<9J_^Iyp\GC]jk=]6: f =@zO!%V  /~CN=?WS/`"+=nq)sirw3<zܧϼ%s#ac[}fD^ݽsrGIlMЦᑇI_16vwqnz\_g4~}F[vsþN۽{oD s8t;c涖[ }~hS[^xXuGR#6t ~a4IH 3bUk .D;aD>vɂonOy]`!7zUΦW?Y ҨBGN jP$H6@ !yoHR`6$dE'r>대 L*&ZWl5=)pNpGr`?ңz*U m#v&j4ldth8wbGy盳<[2~H-RsbtQt@ @aG 4SshjNECfl4@5s@ QDb-L.%a)] ̑<:Lj[6u_ H[XJL20ŢRr5Y'til8!#tOҿgᲫߚzŘ rX cD10 YmCU YQz; Y(tG;/EnOK7 ]϶ERSˎ$P.L-즒#?ͯ97Pq$uH.vy5>ұk0K#Yp&xBmjh u 2db.E(lH[%C␣H6یxc{dhsv6Q᳓o%ez7ZtNz¶-jN Io3o?UZ00z{pʝa{'ص8ؼg{e>ZݟSlXy-NA2],J2-֚Ìx/F!;!H=Kr_|s1lͼDlo,fsdbPYBSDJbV5$pަـфĜ՛7&^d$ Y?{Ƒ¿lQ]mfXgB?e"Tb!)9F3쮙+0?H6wXHJ:&Bh,Ѳvّz9;yCs9) CEKkƿu ? 5͗K,/'S ˍ-&w9$%O ^h=wl3y`]!2Jz!dZ[t u2*neb,u $H I֠0unrpт* ~҈Hh}=T E,S"wYu'&J"s!_Eѕx)2%/Ap!yC+P1B4}-M(Rhh-ƔNRhSXntfH@` K1mHZ86Wv1 (<@0/+d2$ER̃1qT̩ >N( n0N_'hy5(l7oNf7+aQ[Жc:tuh篭նa>߮}$Ӕ&KtM x݃_0ߛ+?}nuѿ%x œOq >j+i? ϲ/?_ n|4~ Iڞ@nn>{i6i|Q}z Gŋ^9<6=5k{ls\g^%֜NSthbJ_.DZOk>h7FmhTӜGp\G?;_ޟr#k'z.Z̓E /~~@צEa{XO=+|5V؊!aX.Iaէ4bϋ#BXH}φe1/zqA6٤ 0o~0ZP씆N^0OXc6/υDavu[yUv\Gs ُyF_@/nz&ooV|h Õԝ,?~yWvCd%r7?jc9vR"oo`?l c^Pvt˱sg.f\c~Ki~ 6f!*~76br>&.%l`,X,pU>_8!qFD$h"bF'bA(`5d$ FwZOksQ>;MFe{JzQ,fHz u%y1('~Z>[SV 28Ҝ C}{OsXfΡaGW"X5!^Srk 8SeKgұva%@$8z-r.y=>^=~#eS&P(9ɐ"xcHVr b**1!2(}%VPk8?Wo6in\W޿9 U=ίFtdJxs*P+AT9lm=g(Õrtn6'c/W RlaJCKI,Gn9yOԢ5Yx[ƇNcK?M>כkv'SnRbwhpKURW]fj&mfevKUx/.gu~8^JV%Fњ2E ^JJbП \mQ]5wXmdk:c[j^on;7χņWW){#Sd(d)"7[B>|^uůF4ĊF:(aC2yaq'4u4@R C&pk^ctv43/wȺސ+<ӗJ%d],5ρApD1qXJv41".M"])M_A%r ,9sUXUHTo^F2~=+p5$0 S[A5\r/ u:&Ys05Fz;w47_^!Lc#GdI`%'ѢX"]$/JDvTyaE`-\qQ"i9뼹*RAzsrZ1$q'dW}U{t͕Jcr{^2WE\bHZ;u_z̕qkES'.dZ%]~UR>.Sj$\>^:]0Y;fggQ;>J%3V|N׉˾>Uv^SwZ}v4U?Oo\e|ГEFͫ]/Ԗa+߻+R,P}4HoFe:hрLYMzPLnDw;8 +7*FNOSx?r!Z&xܪìru~zCp^L%=c 6 #*\TR\Fc>8W8Yd3%DyLKdMN wJ,49k}oFcx [$e>Y!H`97^'Ka$\یbm=[.ѭVVh}f| UXV:D䲕^/o,b_rɉ bAZY% cR2s䏉!sAk)B^s[܊Pͤ<4׳2PU$cH:U2Z Ję@8CzXv)e&ra"1}Y!0)'ɜ-Ykg^} l0QrW 6Xttw1D d&4 ]`\y@grT wi'Nҙ->:J2fi;LJ:JM I[=j j٨z?i̙pѕB;-"hFMH**<0CƂV'^hi]]>lIi/s]i5lJ5#VW!v2MXaY+YFm.,<Fd@]8102&iyK yCmEDHIJriW)X僖oQ*mM4уyB|p?VCsNN쒏˜kWc Ɩ^͐jz&xQAЫiYX/JT-'fгlbdy@ (ʧlr9-$hWDU@N%+`WnRǺXx[x?Oʟ|kɷx"1:25Ki@ۆT##{6&O O5Pw5;UW,P1kgQs$,r0]NX~|ϛl:lZJX~ZvpJɥRf&0Ah0}]9S%\0072 9b@s}`du^HCr0:E-s!2.QH?2cv~P+sknZuD^&~Ai.{!W0{48vx@xUk}R!J@٬؟4N%\YX0U TH.z؟Wroe?9$!!^P5[ .Xhdz,Pv1aVjhoA7ּ{.1g7Nn~'MAYe'@M)O+} {'5You&vY&C0TH'~|9]H*1HUy@b&'xYS*ٻ6$Ug/ƝA`5~J)RKR}R̡(j(Qcr5UGt&ZFHdh0v{yDFt4dr$hW= b\pRs"8]_tEa'Rԟ`d'  <<@lJSN+o@-[˷9Z{x:gy\ww.4  *+IpՂ a\#*c6g?r4*:8:EFU !9G 8e=g/4q,tJ`U Jn)`T9w#cwJgXg슅c,= w]d͸%M0WHwN^?u!= xm|% +4B 'NP40*ġh dW)iƞ pg6T^3Aĉ6`d1芜ۍGa<;vEm1j{ vǻ*;X >0!! ܡ'4c@+2)&!$@ 5C hE{bMK|_u!@<QٮӾtFxX8 hp(슈cD="[&>qN\Śs#eІ D+ABS64LRQqreD= ŭړ]GuE݈xuql렒au%M=.nx--Zf9Ouq.DjIYe (jHD[=.;nx[lYGnAcFQy;n{?>c6cQu<9 MuL.<꼙Z:oR:7X?=T=&i­Epuźd|P6ʗ{U꘣kc4V0)42iN3-4!*d ZTD ,I.F-ZÀ9 ) bTs1be80(3Y`jDz9m"ck{/q|}HbHA44'ˆ/_zٻ35̷u3^N6͞6l8yă<1EF9Q' BiɊ u [wno(nY#TcFx uĤhJ;V&og-k@',UX<{Շn:ZoSbvEDaQ[Qd(J"5RS}rҁedNxҥӄ.EHoԎlG޻P\~rx8x`Ԁ0Ϣ^Fs$˵>QJ= pFe}o7n6CRw pZ\Sټf Ӡ S(6}.5'T<. 9G-`$9J8nD)yJ,QKu$r#^!r=07.IҡJDs9IqK"+),N*H\f~ }AcjR Np\rrT@ dZu:"Y +(ߴD4yI/ӻE5WyVaMvi F-3,9ӯxD"*›DFoŏ/12x0JYL)z}:P V|ZdD>-VLbNl|~=vx~j3:!L/d|VL'ӣO˹y( c?H[]pxS; QDRK4-t3gK;\?"bu2Vǁm1lqrjl)ٹ{Ml=cmPCl1kW52UhREKL-\ʱy֑-rBMںl;r7 N[LLKsmsy]3E} IOAg^qMX[r5uƁSVP^#N01STNLBc-n1պ{f}ߛ`cN&igC-Դȷ ۄ턷1oӚMt*C{ay XuqobxDtqZПz|+Mѫ+sS5*%OD\(4ƘHjIF6j'MrB֟jtr8==0/yttu7<]II|ybd[KDcT䂵9)12ؔl# e餜7{ynƒǪCGoJ9 h>Dw}<,=qr3=4D[rAb "&Frgs抗rl16)K]*(T0NG{f>NQ%D(aGEB%ȕXΤJKg|Az-D:υFWQ#ExiY2D&ID!,ܝe|yvWj./2%_֊3?_L|lK<;tY_i+Qo:?,J7E[H4:ː,Q+UJ&IPJ*\?f$=Pgq'f$8^Ls[:ϨH<iIETQ,$m@x %"Z(@m"qNJqO(Y\ѿu aj!Vkb|d$-zR%0Sye{^|폵5m#H)ikn4HJ12CL*#0Kb^¢4\Qԭ r*:nQz0* BYW["ɥ+LAY}CaQN?V (2Nsalx- HCupjC#c-&_Ax]9PmOԴ. ?V@OU^_N/,oaGPZ)94*Kn>wY:(FO F'oNh18zTX;# ڇmC0q4xg a<^'zz3l0rFQ/6jŨ|?6>O|b|zLh1ϼ9jS<{`T?.OǷo~x6ǿۻ߾H||5ٟpf#L¯"}=V] ͇Z[7͸WnPB2~W⧋ϯG=[u/3G{('g"RԱu-Mz pCje~$;S$a#17j" W˛ҟñOuAf}"׫,~'nz'nj9B'U *:w7D1Wˋ'76nK RFWX4]/_,ig/4otFpo* ?$vʿuk˛- D 2AGB@0 <>' C<8p-!($ 7 Z{gquͮJIF'4d];Amu;m.ݶ{c{с;B-[ntvZWPOe fOPkxb#u&RA8)W4"eָ~mEj']zPYLZ?D`)QUٷ5B .SNt$q^x#&NƗqjWoBÆn=}ЀL9cg,>aBs^i O&"aOٍ6 3ܼMuرؾi0jϬb+f~7Ocp"}$*'cZf#G1@#Ǔ*Qަp `}Ջ 睝H89Ol<~J~ه.&a.ُ_~A K*dAj6$+ a/10b R <SOr"ŜMB*n3eb9̩s˰!diYw V+c~{@XY*]ؕ6JPЕ54rg2p P)1ZZ®zjݵ"va6ۂWqT̰]Cnb8oo߼-}6>W+L:GV^f+&ok;݂kkߓn[G8+ϻt[҅]*wʹ-[eD)(C~RܸM-w+́$I,urDI Ѓ+SMsa!s\K9' =Mܛ 8.$/#aj ٻFr$W?u/:t'!1 1P(LqLb>EMv.[*rUxSʽTCnDSh6$N.GwRJ>I [%"L%Y;:vLEa^@מ? 7[͇Aq >o1 0nҮҼ$Pz fV6+pVWd|x0m﫺nuW}S X(''cG&'~ ڨF:Z幉I82 {n}+7h4=KصOͶgZ/Ĵ,#C٘H U(v1ka4VPh+ҕ@Db,^=:bc:c::F &cRK*+k02fPR2y Y{|ɉKd=E*ŅI/,+ˑu X[D Yxo0ڔ/]:gC:=4vikǗNkDixqQ ?7 ~(}lx6P$ԣh7 \2;>~ǗftV HP1geD֦*{'=9|prNꄜ|UGNzB4Z稴(so`D`AF 0m.ԩjRr)i2mJ,,Fk29rYBwlv&NKO6[|0V~8vͷl5k 8}ޢ-}s{oyXaw:vmvhz ozr3q.?k-Vf!n'Ս+vpĠ?!DT8ۛv4z̎eC}4RZnG-=Ң'dէ=fn(uǸ_ptdvyɵ^oq޺]m$\3<["|Mr%:Cu˭?J|DŹ\TrGT\gsXU\ZQbr?J|3Glv|n>H\~<{UVUYWo kpC* \ղOklfn' u{&g0NG#&pQPʮjfܚooQHuF0MKq>0]<.Қ"=LA6k~*?͇Y=탙G?z{ܸe@-Nj9u'G2R) T(_/`/ޯ{iMұ:GQnj# $Kqa]rPlg~> onKb,jJu;*-o?xaQH mY4X|,":KF?'/Rޢy '•~?"qQ!+V qpERٟ2-޼c%/]IZzOVPyIO pknMʴ:ifp㦳,Zt]y[gk5?Vuk63ë?v}4%_Xyd9W\'fmǻpuΥnUU? 9,p/_*Q} Q#b5%1e`(~% ER뤣 "F2agAþʹ#?dP>*ovw/ 6퇋ZmC.~vsG>x$[z8n~$(uLd/1ԱL,Qpg@hDo (! _YS>xKC%_3ΝAK'滺ka/i~uч?s~OO{m_i h:"6_F{>q8ڷ Dv! `i$ɉ+sv %gnrrɀd0v{yXѰ":!D$K͗Za+qv+jgDK??=y~6֗T(^1NVǛTXCeO]̧'!=V*ȳ YJa8*UJ6fY ::ٍQ?aLO;[""v#b[a91yej ^cb*rE(v ^KXR.2l㤀Lì рg|gI'M$atC(ٍ:Fٍ_M'\Yggd_\qm+c MN\WUzsd"qQs%BȎrFPb7x*x=:,v௪6'r ;CE\s=DҊ*4)sC,evYۅkeɻ ~,Iy _.b!MCWʆ`*L%C9JX>VDcQ9h3bAKȘ1T(ULLp׆ie$׀c٧dbh@PR}Ll3V "9cd:2t&NKwSTG qy3~^ ugwgjXm՛F~+i~xf1myR{1"0e읃 A}yM0JHXN vs)3fosY>0x3_=݂0-UeǯC|:Sdzõ 1d$h4^2ɭV%s\ 2wPϒ4="< "\=0QZm Yr'J LLт,%$c QHF+S`ũߤ"&FcPR{s))B:Ef)!9!4D4eJ?ψɇݪZPs[Lfj-ns ̙rQ2L`3Z D"D02D̲r@쵾3_r+feO-t-82K'-- =X Bvėˇ7ATIj=?<M>Foi_~旴gqL?2}]8Vϊ/eva>~k}/6@OӢ%iʣTSW?擺GwiDW!$0||ftq0^}Z-WG7CE.o>\]QGjzY [FqDBN _HGXLK W7 4zh+ڽ0o^h7m׶]dzy9;8ۂE8 MMjsڢJؗڱkqSCV JZۤDMrNmԶNil>yt\kO}#jxܥ'xdygew䞻~Gx| SſpXzԷ=ٽ?pcf"ڽ]r޳>xbCyVu.۴npS 4 EƟ -sXta!NԮQvӸa4Zp\n9{hNraK9O\<]YPPJ;0*#b I9I.吼!hŝ~.1T3_JF`W/<_/n J%dz3+ITJv=xqtc$7hC_IŰY*Rl@ǒf;P/HO <@ W!"diGcV1^)}C *^ *3L#?ZtFlTx.ŋJ8hmI-j7\ߵ^[r2g$OkW@ތkYQ@&ɏ傘|'=105RΑ@=Y0{&a5G 'iFӫFϮL稂uȮQ;ΊJ߶YV>^5ףiMGF?~ܩek:6'Q#p/?~o_}W}~˫|-|o6 #]8;`{M=&O54ZZn WL]p_][>7F?_yY!MY hV|5~8$_a;xjzaśbљ-c5?3 o *koTJKKDD\K}HȁYKrrě[0Nghi W!.}8qHF$$h]1x:UII iPJ FlCXm7mpG鶵 >6wZ;n-8BW P%i޻#ԝrt? =8FaV 2m+ RJs& N _[mYɇk/Œ{kd2 JD&HBsɕ, Rz/^N~X w{ \Yx(//됽gr>|<,~\ &#&[ϜωV獵d L!3q?roQZرؾM۵br'+XwuneB X4>/lA4~2,MxK^15A mmU݂#IZUQ0UHR9U%S֢^GJJVRo5/"_*9X )!HFet‘ )G$in9Dez6ED]YdԵ::7N83ZYJ$&{6E6ͦcc0C}tw\(MZ>`y mM!d=Hkb+>fA S):%Z$eSA&g} ӒnLm# 6W( %罖+!:v'WƆmx0l3=bš}v*qS=x?,wƫC J"*-O K) )!{A<=$]`j8p~kඖw+u#Nޛ7ɄxgK7& ɉ>42h0.X;] s {z&X2u!gЀbp.ifLb:9I(<ݷo}~}og]!7|?&8d\J~ @jNG_[F5"fZZ/1< E(Ӊz%Ի*/1,TQLM:A܇AT:LS=XfҙhHh F' zTY9c1meK /;$ԫH*ŅGI$ E!2pi|4?88?u@>'= sRZs+ymLbjzB26稉;40m8ɐ^9@2,w)!E],\d3˜; {{Ζn ?/ꄝΦP Wٽ"lqj7o2u#l솮[*}ygO:}$Nm6ﱶBW0B 4 x,z)xLUz}نNm6_uJ-nƛ;ox[r=<\Lk[]pm`~:M\#7_qx֔!ZqT+^˱M鮁{C AߟNC. [ۡH髒 ?lF{ByVhgLWgfYYj!<+_`cwU"-'5j) +N1tTR\&Sj}9?up*6q@*61ޭS g$=*aJeFt&*vY_X e&1c^i9hR hc an #9o} 5ZrR%8uR]@>djfCM*;RIaUfˤ ^S.r탌IS0&u13u)@h)BanE\Eq0x#gfeI9e :"ȒVho 0! i\4)ge/B; [(&DY ƸQdeh%NgMgK<;bUz9k^JbEǙN^uvtP (R,E#*&'5@(!ۋz{#ؽ)Kٱ))L]!1) ʩd)P0,> V7V+O9i0$~%Q30VEG 5,re:t̃qP5"{ϤFd7 oCI+lX,I.',djP*A8T ٻFvW<%zK/fAaI^6h$9~-cLf5EVw_>` N{lcg#1qx/ yEZΐW1 y,CC^ck y!15Ɛ415=k y!15ƐC^ck~Rck y!15m!15ƐC^ckqcM,WZXy,}rd:k:MX&YgLu^D_ K9+y朁;_j@C2Xf̚C4Ry-\ t0&%B+W}cqF1<2HSg5pBY}ҰNUZzaq٣;Zd_Lbo?je' ^jűAGȴ2 szHof 6UK2 ;!B@/T/rɥҎqEܰD]iBxR2ҕ'O"9HJLȲLZ`DBJtAZ"ˈJ6W X[! 5Zi&sX)gD ; ]~ymVBF*Ѵ^|4FSrC_tbOyݶnݙ 58LB6reM4 uSN1b7!&:$Z*VVJa d̓PFsJT*qu1tyI 0xmJ|Zcp[8!\]Jgѫ[uV4Q=!*DuV~scv!aCFCS<<2Gsv֫%MN0V6 9u7i+U4"))HCxR-$JcS̄ڛrcK ?[}TvHyrVvCT쐻.6o=@Go-NsrSA k}?\\YsV5g X( G(t9E WJ|C@[/EU˨1fAGYrA'Цc^DNQq%.ms2*laq-TmjEmD6_ 2(p^ \uml~ϗg |eX h"a2aDڳsNkZ wz*8"4,dbdLgi,W&{ RZ_, Dӧ\E 8!k2j=YRI:I&RAZ:!'k3EJ=uSIU՚%E[.f]<6r Iu>EE `"?ʙq23.@/ab͎3`uᮿ'0aw?֛_#GL~:n(lIp-{GMhS?{9ypGs)mGsGs r^ Yۅ+zn&<<:2^ noEZA6D;)CgCНIC V+j'Z+5栬$l  6BE& +͔2ӫS.ci0CʒS9D[ BeZ8p 32q sl0@5qFfm -5~R _)HNvK'!}\mּ/W iH1.򜵱LPh!By2B&rn:w}Uru% $0%}obrrdJ@2;Δ[M|gzZ.7yڻկV{,uاĆ{@̰뢻(ӏFHe?2*'Nx#;t0#Wiȟԏ)PwXv$fכٗE4\-8յkz}?v#0vA eP)B_[ѕQ ~s(M tYf׏s5X|o>] qB~{Iܬz~r!F2{D $ ?jELrkń9QQN%?Գ$u/b\q0LDi})TgфV 21 `LE! kR~2< fYۨH)+H<מ# 3G?% 'd;թ湼үkBjqnzA-wS}X"8rS s?1"Gy !BiY\Ũ"2MdؘQYE3:; IgwO H R #^xe6?zʥXGKKHoGDML%Ev^ĨA&'?lc@ D{#]NF& -czciIHM>Fwi_|4׋\}NEuDZ. مrp`[_׌{Pz<Ξ Zuݾ;#9?n2<%"%z~#fM ޯ8".=6/w p2k"(lKh>AtwcYib%::wߎ xܩ)-BNݓ{nz萂Ziyss+)2j 3ģ K L DXҧ>+e>T ϴOD`JViIc E>%asLx<%m>:WȭUZHWf`ȗ:mLQx ŅUSz{Fv6d:rq귿 )#}g2$';/gB/_*}9=`wVi'sF3"j)(E:fqN}YqIg)V;Lɫ6R_/w{j[k#+ 6"ɹL[GW.GJ3ILfĄl9uw~%7fw7= ?so~jcqП9b{m7š'ToH#y{6GCM: @]OT@y8Go,$N9:qt_`GL68az/gU&\RlR2LEIz!& e XJk9<̼P{4~ -[yfeY*-fjW[Zbfߦ>PЫ߾Y|Z,o"j9>Hcȡ#]ط|2z,jejxMzl4֯ۿ?8 f=^'b~"۩r}l7boP ;:,<"z7if$%``V>(R(=: 6N"K`e-|Pv@_A9e3{QvsY j3~ޜMkU̔4O謈2>m!"Hzr- Vx`ʓ'kj]JZi|ޙ,1ZKը(.$i *Fڵ&mli"l4sk@Nrڤ\̜20#8b 2xIkTWIWp^:ޞ`^Gd2$EB|AkA0+m GȏwE7 Z~ìL'V 65嗋盋ueAYna¨V4QpZe ٠+2MiߗS.vg%L?᏿l{_6Wx$ ޵6q#ۿMݱhR뵳7[[Nv?$.2cIk _H/ Hݍst׽0'yߙe&Y͛?k l .57lE3bP}| ~Amv/WWݭյc5(VH(OfnX`ѿ R'ї3hΌ~M57&~mn7w|w~߼헟_2}/oٟ Lχk݅>?7?ߣiVެiasӚ6G=}۴vo%dI׎RϷZ8~x2uVUA0tфC<~W: Uk;~zսzI҄P$}a:cFm?o]ۈ&CENMQ}m]jWͿ*^j> ո{?T)*(!n5ž7W.K|Ս{s5[௷&Be 瓌|o\,~MGNG󿗳E6i8[ބGm9![@ :Z"Z`4נtF'8(־1F./ qI8H`Yhsc 9oc]9jppZ[ @amɇydzmA4fUIW`b%lM+5GY*𷓞80z%$$dB%!*amIw$B)CN31Z q!(\Xϱ@N%eL* <#:j>1TfWReougs:&t\ҝ=]CKn0 6Vo9Z7}.zY ق㌤ "T*@(] JA1Ib` plrK LAHIk`(ǎmYpA锼Dcuc"RQ R=mR$v]9T,)K$)P5rLl֝- ?磯uњԘJɎ2.fsLfm ߵϷ|fNژ!T3[mví '=]n$;UC  (?>enUjnl;NOug|ۻ;/.nov+^>Lx6][A?mqh˦lrswQK# p4gß]mzmwVKϧChspwT;d/}s+|sz ͢l-Au4V2"%#[.GCγG6Qss\K9'JE{F vG-,:P-c{L@'x|&ilVC}$&z|hR l֝`X3)sDkCp(JM&/Z7<S,@ h6 rLž(m"Err<m1i'KVD 3ё`| uc+,ԛ5Z(Gbs`SX9TZ ".VdX 6@uFP@^>XF8g f|c2RLiK-\ɹF05l ymSF)/dz[amVJ8@@!Mr" lۯ";l% a$RWaEɨ]ENE]ao59zu{ɊzHP\k&NF] s**VcWWKC )QWz˩/ o.Qw_&v|ߖ9=ۏ $|c:/͢Wnp!m?y~. _,7|ǃ~A=bLR:ԋ~2hi~yc8^z5@}/ՙr4\vVv/2A\ɨlG @2,޸13G>]Tc#ù2Yg] u.|օϺY`L`bϺB[ Al!-,| pMB[!-E Al!- B[b Al!- B[b Al!- 2Dt>քnݯ{-[QuJ~ҭR*(ҥtSSU\CV[}}ZH_IU8$]2VU4"}}eV+JP8Rr& :.\I H@oaoI^U!2 :hf8W 82Q1xΒj}ۼugW:q4|\| ?aJGj$JΕ N3.7Ø}&ڪBʧ2 rF%Ʀ V*:f>Lkb|wl>Eo::^~z/_!_DYWط%sp-SziNt5pEskEv>>ZD :ʯ'HyiT\@IWNGR*Dћ =9Kl1`Q./&+9'B/(X_D/r^O&b!e3WFun /X~U?L/[H6]ԩ gI *Q8DR Q;*pZiXnQoosgi4Z⎹\O&ě7݁òi6-x멼0ojn{?C 0<ɦUu< zmL|l$zy@4|rIhB'HFg!u0&v#o%Zt|")'b΋2l5?vCVltSCj7#;!bgjRMʍ_ݾmVNVwXTbhƯyiKLnm_ }/۾pnm_ }/PDqFD/۾pnm_N }/۾nnm_ }m_C8}е}|]?W6S. k#ȝepsX#+tS sl ڮ/NqW[칠 1PKwQW$с>ipT,5 {r?z#N[hlu KKr޲5sRϺI^aʢZŇL0k9gW9XEv=^0z%9"9"9dB%!*amIw$B)CN1Z=tBPƹc$!\T&l@xFH4mS{X ֋߶WEwCyGk9\|̅lz?wE-~p   &N &dHfEy:鉶6sCu)_`[`=oÏL)uv gb]H:6\Bq5J=Vr}"V\pHRY8ݒƦudj[:HJ+C3Wz՚py,(uNհ*;84ehmq E5yr>3 ϔ^@#"hQ,pPBh>1#ڔU'{YV*|QW_k5'­.-mi<2R 2e:p*J\yq|s."mEDoWXlDEZ[">hYft![MZq@kSy/ ASNF8)N?gcnΎA:Z[!#xVAЫqX3K9J \0KH(iIgRNOd hK-$Uh. 8 Q*5"Vڬ5atP[H<-!Ooqz#tz'qf+-&ՁL*MZ㡒'gx[UX9΢l5WxeAi\b5+cNX~`v(aYf A&%A)jfM2=QZ[&̷le&PE/P፧:?{WF/E w{sX l$9g-ɑ$KNmuf=|vʪ 3ʡSVra5sv/.qOY__ݬ(gFJؐPK˓ D(ㆇ= K'ה>> YRrs3=P:( 6j"M9ڄV2dSg-sv#\5Ԯ6:Em]6}\ȥ4qN6ONw4kY)FLUz򳤒"vLp0pғck?]$zԞpq,W[kCqVE;​FQI#wBd#q`9zH(Q *t3)( ff\O}j@<..{2:ftΑ{k>gEEG2ڢ$7ʅ Be٩fQ$f2={ytjr\OB$ QƹRGDtcW2{,kC ػ=G{e/ 9(1I%0l76\EJ3e"5:}J͙yJ/lΞaSBs3|CyxV/nF|}67Ļw=v6ڻ q9{2OJϯgTS䋐6V{Jɔ$cMv2* h"(G 4J@Y% ,brr%KgdcߡT3g?ʳ?N&ܕث{.AX(n]1(R5v~9.zRv;h殣|]U&=8H(eD&-#ѐsx"B)|ZKWB#FAM=aX 8Z@;)IHx{X+L2 yV1ikϒ"X['`0QXm 4Yr4'JHL!т,Ľ1䌩ȅZI*0,T%w"D,EnT\s Nr;ڃ!+C5sHU;[hY˺8#&.(79/=Ztfw[DGe[.c\lVoxd%H͔ 6C 1bi(YQ.bTQ2{Oؘxd:脞y5^!q<()F:В!Z/ Bea$kA z@*S 2 +8{ax^o6LBpц^ECt2ȤeAqj^ea}pgݤ>/A eԓHQmewwH'#+ ^xy5g%[_r)+*22Ñ>a?[҈RyE2zqiAQ*o9Ƴ$p99CKLD鉁ȕ1^`\G8A<;ΓUEy2e]MdވiXaӢiT6u>)QFtGWI'p7G6QNFuU0M{Rqga\2}(+\8^ednsʓolF2ԑiCsjэB3{uy36}|~JJ;yO⯻'8ݸا}\r>|isf~|3ΧդM j%՚}jr?ۺ]^v9QW3wtաIw6i|wBk~,OjQ Uxl) '\:OyYE%GXx,rxost^~~,W"PW{~iT7=nͶ#Yr~6,>>}WL>kX; j0ܲoծG+5=GFW[< `O0<)}ϖu68M,OYZd˟2eAӤrcf/Nlrp}kgnNhs]jlZ ZxbyV)Y:n{G蜸.[,l1L"ZY-VcM!11CJ:ژ}t{$X  u@ϛY%dK兽Ǫ>Sz<J%ϔ/]ɾ\rQr6f:){`XۏݤwlT<ߵCmv6. ?O}^u灆g=zq}N1QG'ML`# hl^X`U@ &R %\ͯiKQH9"$lWQ?w !et?>Gi2P S\>cI;mh< Qv2!:cD&kBIN+!7#QNe!z\C::*WV?!T#`*bmIl+D$g"d)4.X(`];'* dFNomTDlTXwLykeR?yscr3oHov)عm~[~g~?]j*5 Y99imgGw| E"9ɴ 漌`jGA›?<,NMgzVJWQ[!Nn~)&thӿg"ft֮rwcakzX-EkbMȧ@`Y;ms[K޼[p7X1!FO4\@婛|M-XEұ/2hލaKhF߭Vjut(otF--6]qѹ&/oގh4HޛJ#t$?zc%쬝?'%bLF`C!M55{hg&*j. m0Tzi0ԉ$ZIL:a} ETbLV;KukK \hKf)}!6#Y!YTw{]rhȵ_dU3g%n6OjWEnLϭq|H iX/cP/:ll>[=3=9hrоAEcYP>3O3uy=♺<3L PӄVLd$J̒icePy'fmlG(`U<2:#:fCGȵ+2g }-.AW7+v=!뇷Gƒg64<7UA*>PTUo k]_DWsQx"jJTS=լGO D4DFUn1EO}@Ʒ3fį(1q^ZZwK 'lh0SGG::֧ն؜^u_~^yZǀ問h/oձ %rJ>dv?H~ ҏ&Xf,btZ+ Ҹ !*AݒtG+d_J+u6:aaapIrÍ23UJV$h g!T}AΏP`(T+H'-csZ 'Rh6\ . MLUc/9uC/CʣZk_-eCcG'tѾxrhwӫd<|xCeRohd1dÄ%6r^ai{[K1M2shp( :-XvBse@9AҌg-nxaDJ(`Ί~V͜=߿7[%/2PKg\ f`*:}PDwO<kmH@_ξH~0ć &8 IIW3"m]]z Xh A0L=D\ p| GzE~Q hW{) V1U-ƃ4I+=v2ZJH)븐.pGNq2nV3CESxi3*`=4ˏ9G9X)ckׇ"8d[{/b6Vg']A:7Y8̫^H$WClL&R^h9|Jӝ'f9V/fqtK5 fZKj1RT>/r!b3P$LN;Y%:߆I8߽-}6o'&դGJ?T=%Kty9W W؊n]BɺycxѭtPItJ @p@.ESIe|3Q)| L4bC%ܚfzXi7aLϼ]/I L`@$^~Gn6vV7k˭s2f/}& Q+nq]w4ݷMFKCP'~Zq}fZl[3go-7I5n7ɲHXNAiUЛfUd uPvzmAX~[dyf%_Y)'`[?mkɯ5^(fۗ%ylWs?:I[l4܋1iP)a,O  5{-t]B}4.Ȥt)$N`91`}F[*&V{LȪۼ:;8fY~7;"{dFiex=%{/Y'JV⢨>NNS‘@Y:"s9HHx<KLJxbS`m))B}R飶ěU`̙R9둱R qƶX( cpX("i=i-ڇRt}zy j37Wǯۡ8Li1ck`"",ǒaOprAD[/^u(!{D%$ؤB:`^%(M) Ӂi]`$WYض&桠v1eaԖjcosgfՄ;O8W4P p)oiRm Ś2S#8 Z 4Cv4H$(: c}9043=#g=f~Y" b1UaDT"33D èGH*Z`pejlB1* LEq L($^2XoA΂$LR#1sɷtRGīKmu'\l0.\|0+鉡4Aq && N(XD86>pPwlî' lۓzy(i_®Ev n~|&G/`sncWnC̚ɝϽ ͤ>cWWD ykcR1<.] VhY: !@o -RA8AYTQ9` -R2 AV$)QBo74t6숸ex}ڇfVbqֶ٨tytr>}DX="Q*-JsfAqaa^plcyYH1l-Xhm An`!&PB1*lJ1rKzg dzYľ,ҫg| soW 3Y3[夾 Hi; CseM5QV!ZX$`G~v$?d;ɮΚ8#د?{Ěq1̑T++ cS,&c{)`7nΛ!)[/c<=]9tl=}7yWM7{?S>]8 Ǵ`. `jZ(֚jxcN C- TDH|hU`Aym``DJ ,5'TQ)a(0`"KV<i0KI % wK9 Q ƕoU0BNO9hli2v@ǻIO<yeL.륧30sw:be9*$q,#U>r%f8 =|nC-ǀ8ʐHך\aIa|T5<w~wƽ+U7R٥rO}}x>o8.m4kHtZsn\'3 3ݿ"b2Z?BG&fٞILΛPԙ7ZMkfodGÉA]=%YƼLPj={ _umfTy&dL걹~d x2=^a'"Jh '6۞t39/UFõ!+(\ڨZ%1+ŝv睎+P"ҟPW_v}Pd)2}p\lDSc,>ZoVp&F3ā.󰸘J^bt)>|a*Ἑ]D0xЀ)]G\nj݉Z~]F%@% 2(PF`r[䬂_b^AGBh< $EoP=$]mմeyzV@TƆ>׫~+'9\\Ӫ́Y\&|_7?|W'3ިJ+h?NN+꺹HΚKYթuK/1:,U\YRyWs~2nh,z;Y~C q2ǩeTRTL?\xƔV^3[zMeg{D6m!AP.ćg{Re>K3D;̇ PgP'pSg>KBqt8Ί~*b ~m_+/YO~}mV9Dh6SHC3KFDoAs*wʏ{!%ˈ,'N%RJ4⃷QG)"o^ l܊3q-F-T3<ת7z eOpŘD]jO J2lX!Rh$F].voyL6|e(a<<,[~&9}j@_kai82L wTtTuz U[ϓROlIkLϳč"Cg[z!*[$reIX.t%UZFH@ZZ1~Jt:t0(dH0HtBE+qpakN  ,- zử5ûY'3<#I@@< G,JX{U4E]$>]>=ʣT,7(7tf X-JFȘnR嵑=H±]s{I=њzŖ3:_E7|-#),(HsRv) J3;1p"H&ҏ|CƽamT 0T#22 ^x#)/N0b\83/MeOnŚ5׊ $ Dc)"23NhDSv.wK ȏVbgY8.\82< gs#7 z ×q)܂NoExV}Z tzu?fBc k[A;i1>ZhOßTA@EL~jLe^?.:& )>QaUhP{C?,\~޾>GLǷ߿7@30-\kQ{^hwޣkV_5Ul{ 6G]Nx~iCw`q(݌ޯ7_ %*ahnM "b^JjC)H&̙#7*80 (D6ţ6AGTAOd,(tQܦ4:2,"@ a%k%}DĖ.Rc ܱO6,vi#e^;D>MO ,~25|]ISᓰk 㮆7XG"k]S쫜pтVf!#YƸTƌdKQieQz[MkcҾJBDυd69)CRn YHHaЪ*C0!B*Caj3aa2b=6MVHKD kk4+h3Uo[a?l1׷۵ӢZ2K&i[sI-%nzt=avJg܈d>Z&6Io^Ť[Rvh?Ί֛t'[l^ͳ}V`7MہJ -7^t|M>J è)1~h]o} J{Oe*Uib{vWwgͪ nX]/V$>LЂR9˥mTap`jOko)bL;elf1Y 4<#QsuGJ9Ŏ<>:7="7RN㹉)| ZKpR$pN'CAD.%R8eC:yYF+t4vCP Dib(OpnwG=҄6zKRU#\DZJÕ60|0[I<"pKvW9B"KXh8ӂ!,6K":X1.$R5>s SL{vXERУ%Vj_!WǚpFVq]`p2BaψT6%oJa Q%w4Rwg8 5:7|#0t5 3]o n Ɛ!u˦Ͻ-M*'fo=B-) )Ov91({+t@HmojRz\TknP BvϠPeyih8eHDxq-ےtv~{o޽{+rjrhjE@6W>_/AU)jd/$C_h@3A!8Me3;:j5[8;r5,KXA' )NQRT [ _OGtctjϜT7sVz 0yR&׽%,2ZN@UI7qRMH@|.g!_x;  pSMd$ O&@Bo a+HXJ ([(G J1n#'ꢍ^)'* 0" ;7}s{.39mAb8 qԉSEa,VS%&LV=#z!Uxd!R 5H= FL ` Xy$RD[Xk ck\mΔ)z:ɐP6(V6__u-􂽯mT^Lޖyא0{ ZlcZ0HKᒃ&Xky1:9q$b3;PQm0<aPj5ܑ``C "!4 @,2Yj# ORpLKJdc>":3p8@Νb΂wDðqmgn ah=& ;IҒ~$.pZlTQJRvaP5~Β?>0sw:be9*QJQ\^xV)C7"JȅVt3O/}'|Ԓ<7C-@q!x6HA A*BTqP~D%;(O+' ҾXHqȔɟz,sM ^Ƃ4#˙{A!+J2|w[;GtD>?ˢE͙m*!8$ahRv;UH$/A.\^eB($ZhRj_0j5\=A'KFڂ:2~2Ki9aRUONyN$.F㦣mYZD ݑ#7q;W5Qw< / >/a|a&+K3 k? ?"/)p+_1gF͠#u<%&0ܬ gkJIo҇"ȯ@MK:A[z3^\bƥC\E4t36O\f˸f6 j6@@DQ;k_rq =Psɣ:ΗH7*p)h+{J3[lU<.ҹ]Ǐuސ/8.me d0$ҨFjå/*xMU>|v3O' m*k˭[qVz\Rzv(DT'5ԨH53`7;%$LmRViI*6iRi׼'$ڰht]-6U3/N73y!lFoe9J*0K Oxlt3W{4w'.xH$ulqOG<%aNS qRJHS/ [15+CY@%lPZ\`'v_x [ ik83=X<<|﹕6ժ})1jfp/߾B۔;Ž,7>v>=ɹoD3k8gU7#4>OڝPt}5_?ƁEɦE+|sN4K=h"&ឣ/q@fjv{ao8+y_9@Z_OǮ"_R@Gcns?Ъ ߕѬ[DTnCq}E͛\?Y}0B%u7ҘS,i.h05zy3z$#Ngﹳ7ՐOxP" `I?z@5޳^BXtwo0m~T驻57O0LgFP Z}7̇DOhq<+݉?߹i1t~9N1rr?]h5 c`\-ʾI {[⑤dFyvAS˻u%ds u}y`~Tz_/uj7RXޜ+~Vˀ&bTeAѤZb-)kHѢi%sy^P3VʥoMCLyejapGȞIqT{ݹQ $R+"O %`F7G6c4SZ)y22ץ Z]-/၌Hf~zxe2: ym>`G9Pe)<7R ;\ta?!'ƨX+wNl1CKB@;W艠O|ZdqěFXE+Ph$T,Y$g\{J|v]Ъ>2ιRb7)FZL BxݴEV[K{NG [1P+:X!\݊@ZrȅWP %%#Hup-R$dH1R,#z Fb<Һh, E9k QuiFnsӵf Y;u.h-%$::d{|Q]]0|p5^*~ F Ԫĭ`)R 8<=Z*wp AɁka 8GLv)m!R%򐫜4ٔxd.M V[ŒL.EɸдЙ/#[0SPO* A6XvQlB5lm:a  `b¹ TV{{[Ŋ ̶p ]H׀Lv"@6$p2Ȅ>%nAJ$_Vb( 2K*=@RT KiC`w0c¶YoY~e+ c804ƎAA. D;1,pPRH 3&rJAt2>*A`) )Vz! Sќ΁Q.0QQ69nYڲ2 K(C 7: yJ AC䣺JR/]G{ۊt͕9c68Ep+ ʨlR; '! py\`oۋ~~{\?o-+15èuqcxp!G 2,. dXH3(Ԇ@6Ue Ó4HI(v 5!o⋖ ȋ7-k^ "5*o02ȇI4PHYvGkq`thͧ1 m(;P c}BLtסMB/U;vB m;d3#n$qȝL6h3zc'NP'YԼjsiP9 ^[{@#$UdO@jw1^,9І(T.[I]\T:*0" t ((|Q{OH* rk% EK :7cCZtPB TH-7 &iK9#WF@<ߐH "sR%Y%!07Ƴ!hI臘ңgo__#h8;X˴ubnʤrjM<@xp5yYG.Z/imr0 ])̢Ѭ\kR|V!ΫG5[6 hKiF5e@aPBOG&/բ z+F-ݗ A*g3{4EC H``EQ'^) V)yq2*4OWu_Kitrp=m 9F_:+ِ @,2h #i'b6@UW @:CcYSsygC;抶R6WWXxY8s$> s mԹ%r@\io.ٷ1j[d~<  ,Rĝl_b.`JsU<."iΖ!B]`-33|bh[N aڂ! 1er(-4E'e*JsU޲[È_\P{Yv<b;j ׹W:-fb ln'us]Jqc2x88j DP`3ug) @AhwUyUZpãosAeTDň(;"xՐ\غowm{GÀ?dME605P!)?~$I-ˀetUwUwSM{>TOVgB>.T| 9QA[*Ld)!I\=9ߩ/剜k-?6&~mژ-i[N0 P`) #OɚT\´, .֝'ޅ|i_w&Rcع'Nv;jq}]B(ֽ+Xb{hXIs6< _#"GV{ NVfHi05`xBGYDXhA .֝xaAy?eHzb+X)%23``ʧ œIdA/Df_c"3+xGa7:5yA}& `@_}}vrֈYvֹ]3h&0 k7OM6E ɶɱbIqH`D`vHG+T'cUb\\NjS4mr4+K϶g{̈́b47̷Glی+kIiiNΫ#謐YVP3(6'`̟ n&ڡy烱xpc ώƳJB? IrRPЉ$8|}?ˏZzhSHviKz<mܲ ԕfY:?ꘀ7`ax7'77G<S۸X`&ʟM;;zfoΗlGJеԴ *|H "ZB$/i_[W^LU\T dÊevS2_]Xw'eԅ.,L\IF@4MYybAְx9X:)qz)_^XwvCYs 2*~qrO-EǚsM ).|qϴ\]r̝׊ROumk%F[e$ AdL!5Zܖ ao5 :w[}l $5 keBe(.eZn38X|R7]),s!y ;PP1sPjkE27̹5BT;Z;:?TiΓ`&z)I1pYR2=nCQ*] 9؏fӘAmx!GEԉ5 ZsG$N֪ =rx7h} >}+ϳA>~<;;?~T1kRckNeG=Ze$~kiN[t{,_d9Ex__O~l|1;]|_ ;\e:s> z4<9xrvK4YK5;J AH_K$aX UͼNAL͸}x-b܌>/_pipVVkwT{4`w liA|2\5׃q=ȗ kqRC6T@&+6ǃ0 tl~2¢2Fn|E\=jȿ!7 ͯYΈ01] ~[vz\/~[}-Epj5 tx4S/j۴W Q6;B W- No7ns9XD-mť8hdoMt57/.. &Wlt[9[~C0lkB2s!ev dbDL-Fh55 < CZ,՞?Bxx9zT Bp2)w0;e{%2S"dh2~ˡ"C48?dN @wXɄ{nsEtk1k* 4s %04sfa9L3i04sfa9L3i04sfa9L3i04sfa9L3i04sfa9L3i04sfa9L3i3%`' wxβ$' 'Ƚ|rH$\^[[ 0@ n^O/BH)"|ټ6V,kw㵟V& bI(C!ʀx>t ~m|;xߥ8YlD;/JhuMdJҙ`ځs`r 'H)KP9Q,Xw81͇xzxP1N`1Cv2o>{DSp8]m9_"q/aSG#,DmL1s2r*o]| b4- zAkؖc},,,,,,,,,,,,,,,,,,,,,,,,,vzR,4O#,'Y>XC3!PF;ҒJ{Q5K^:-[+\ O\Y >wzatpѴ:zmrl$˃ iwF dK vIkR2;D^< ,M(bk*8FXB؋ècOaib9Nn[t'o.Fi\n/a|1Ӡgj45台ǼQx4\6vr o8yӼowYFJ~@pv7`YpY_bo1w&&4y͂nՕb ?HM;Rxm]0;%*auO ՓçtF*T=G}Gsޗaq5UXYY 10~^'cVHprw$|.E/xWD/?z'4(ңUh|Osы_=x,a2`rj* g/|`8?yyW"H݉.A)PNd-#w:7iqy> ߁_PUkH%?ߤX/s%5Θ{2 00mqѭU8lo"H=G.*ߤ+(P\|l^+ZLIDR/,R"Ļv]1AISW)XaRS \{q"b*!9^־v2Dlє ٨@;U+sp7:XLlHº ir'Nu+3Fj,1.MRHCx:[P+ -QeL9. Y6iɽFv鴡sgme AHHD9>[3h^0LW|b)@NҬ-+qʴIۘ绮2gc`Bhd@ [Oพ/h L$ #v_9(&i6~YR3.k a9ܝ+(^Gtv6.Q-G28ؗNeVL֊Q#O XQK'V|gW &>&~wihֽGv7c1WTuVI,EPKԦҤ(OZyM%Z_]~RM}3UyyCi}1Ci[g&OIh#B$7sv1Oz*U3[E3^"NJHm*KꨙI&4s#jɃEAE QQ+UG2#V"NiSc8}ԦcPϬp(M`Z%hq Zź|m?X'>'4^ɞn7rz7U[oJ:EfT؅wz|>kD_Νص8O=nu!\{3=nPgDnǟKWқLT*t V%+(Y-uj75_^ C͝a2ootswqSypuKѥqS$3 +\YtרBjtV=?=*DV6rX791hsj !qp*g{լ2K:^^|Vv@>+Cz7ZA0JX:SigV;#eMS)IMx :+*A[R;輆Rh* ˰؂="z(hx+ʛ" ;j/*ydfjI*EM%b[U319_K!D 礕Z'KcH*dJ2sy*UL('APTJԳbQZ PVP' L`DEǴ0&%'q&Z(Lm2.*0NQ T$z_,;ӏ t:Ě Z[3LcMj~gCpXWVF9!'J9M 4Za6B? DiOW27A I FFh) 94>YUG+pL|ed 2qD6ѐ`lQQZ e,ag hȧ!]%Q9\D"N $/z#o܉bjz(WLD|-awPZݞl9-IA|pa4gvvGNyLN Q=ʩ[ővj'A/yc n&E*>$i p H`G 6Ls\я̕~B+vqoKN+ނ+"Wݧ>gyr Sj69lԒ+nX%8RTdL8/'i.K-net  i+59Ck )GOg+E<@EMvūx,iUpU3}G%8%ZAJ++XAhCZM @BB'@ʸRMMBö'P"3hcڈ!,3H]Xwp,Cc|iBus-xytۇ ^rD=$*6bKb+܎1h>%v?_5f&pZO0d ,^:d:rw,nn}jjăOƟ>VJHP=M'鿘;feya&E95>gj{ȭ_Nwlv - 6C֍,Ľ$[=lm)cҌ3Ç!Kb6NHNf/w{ɓ"L}f۬Mɱ^vپ4SycҐpm7'뾶e#5-##]bK%i2*_p<'`("  -o>K[5*M[XsCZ窔>9h~ o$72Ufx{"z WiwX޾F[ z'd4XAV2JCP&cG:-Rox>KG-f5m;aEgCR{09ĸk-bv1_>{Գ?l$+L@80Xyftv$$q br$ %@J<{0BtBpȹ2imȉkSH j7c)#V[fR DCp e"1:A>w*Reܚ8;:0n=bgO:m2cRYbk@kRP bPhSpaeڵ{)q9eWDe7MAH]#`&B k%b0Yl-qvp_# pAVQ[c!0<-6f'adbvGYz-<>H۳UZ4X[BabD٦2n9]Cη\Lpsn`02?~GFldI*uA (ǓBC$q1FBhju'$a(dGD\B2h wѐ$uKS&l>mKFøD1[ӎiMڝI41Rς #WZ9 pd9Ji Zd̴Yi1$!eX4dkbby5Ž&1bV!d|uZg;֤~ bkqWD-#Cč{f^dp0 @ n )S2@7{,hrUčD 6 < 1CiM3F;h<8B}7ʛ9Ōa6Y> \k\QM j sMŃV2Q$(Љl#>['_:8q5>N*%SyRuB6`>(m8c/QF;^辣b~ as tqf}z~uz<=H`CQ:-1'AD5`N%3(8;$k1=BMCWI\^ Yx}Fǥ[̼iDtwphNxT(ôB! 2JR\$QG4ؒqbV;ixFI쌏F[CX3:癶*otyȬ  `h CVE: 89~ֻdR 4a`bI=Jb*1ǐFnַ66ֻI0YWM9цPmB(jv2%{ /x~&B54|ʚeCt,咾-Ep!.}XSŵ.{S_/S74n@W$cwðvIy5^vnhjqnq @hpK<7Аv њ\k[ho$Wט+c} ?U{&yOed5uk!CoONJIyD֟Pt9.4ffa:y|fB0ˮ$SjdK<ܓ.`c8Qh~4O,yo]{KectOm[,&d&vɦPꁿт;Dmht;6wݱbT7ufYЂgeB:DU2h_| 21AY2V6'{)=OZ8g(ɲ,H P*CJ 2CJ&C>8w^@$M%Wkq5]ڵ 9fZm^#J%T~R!DҒPd0u0PrV}/\t+JJR'tX,eOՏLA蒯b*gU%Cƒ9å;P/twa=lZJ;TC;h J;@R(G <<YJ$0*)W.ȘlT(%jkQHCKC-1B%W@c2rt,!5qv?DӽF$wK_-)4Z05e_ޞB+ǥw9]v ^J&w_$ I691֙?". f9Vύb|I Rwm^$L ȍNis#dcAz 5fD4#V(R4L3+o]aE: N8e$Oiu%bBV^Zr)ZP(?O.ic-MLw6ekQrFgIrP;ƓVTn";rVZd?zz{B%9 q0LHR$e|ȀV1^_rV7;aRo}Jk-zE}_->Շ/h;lIw\ͩsz/ NJaǫ%rC]Y̿x _0 ť0ki-܇Ywo8r^bx/w$cM=6oI{:m7wxc7yZig~>@OΙ^9!zm|VGSI9yJY:V>4M7ʽr r/7MkZQM:Ycrg'߽yӻ?/ۏo~|O ~G#0MHQso~NCk]/ڴ5ZXkXO=]m>rC[(r 'z}=o,N2fƺ0_?ż0'Yxl?ἉRv hAmϢׅ8fQ>1ǷLo`~0w'_缗cO\y4p4~k,1Y.\ծ^`p-KieL*xD0"EiD.-/=/PTn2f~xqq\\u+7T4-_8(#Ֆx[ o1m]n~TIUCYo͢5W ne6<5XU#UI'Ձ[we[([rJ9^(%3dMx#A0c:9rTmtY`Ե:X3V9ˤ3"2MgZH$(xe^5r-Bt ;v|Ϛ̬1Na>*ACFD> }qb*IETZe $15ѫSI&&l,Zlm}DC$!1 1P^x\`O@V貊8oUHkR5_5O#ZbE+!uN<0踓RIo*!d^elzϑ^Ǯ ݁rJdR7OS0[ZY>&ZQ(c3k 82sS5. xn{n[8!=2wI"|N-줹tEzgr֫襰7n_Z7OwO9=3ӳvH-K1f[7}m;_i7puqD˧O#nǼɞ̀Rt]2Ψor˦z5U/ᔥsix϶JZnm~S`Wj|!HBV7W'߭!i|VMZK,,EĻ?N~J9vٶ0I% &dAj16:a˖[ax!@o"h|&_Q:]qxk^?>X re:i?דOi&]!w 7{?p=y0w}nY`et}0$)RԳ1C׀% r.j.1%}O!"zc>k& DKJ@T*3-tJJuQ\1\ݨgf )D$e>Y!H`97^'Ia$ v$g-rڧ0]%̅Η(6De&#=//zsI+lyU0&u!3tCRH#Vܥ8xS2PU$cH:U2Z scDHc*94B m8B0LdR`2"+F2"%Z$2U#gC>_q[A'6*=Kb˙Nt`3擈df6 +Ljd,Oz9}΄%lYhU1L|dRAڕUQGbMRA8*UjV>NzŌ+~$WH IaPT'T"fHXadGUMj{[t5MiBe4撝oLK,F$M;/z#o0r[=V#г0_--Ae[}=xFMN L a>8J4d9.( h?G+elREZ;d[P{⎣r~5DjҴMV7PgU$ i HLdz&S`)ܕ@v?|B*Vi6ƿmr2~u"T ªӟwfv13r'FM>_$K%Lh鿨>dokQ?Pk*vmkFgsțoyXa0ȅ0jM{J0`z`Ooz7H0J}pܼM`؃` Xr7a`Io_D ^ 9b@:I\8&3B謏@EjN[%XZj<^>%trBus.in,q k ^36b[#<풾خ3Wh6z8 V %c1Q=cq>f,>Š&d^_|v}ҟ2F幛ƀ-YB Z .4ƣMN1hݽذv[؞howȺDJ2:!f22JYP dy'Y >C^F(*Cʒ+A+Dr(4Yef+0ߕ'AJ . xu̍.aMRo{գ %M?bojA  ޯde"5,@J (Ȃe'$c`rvNn K!LX I0R(b$)rxs㹂wv~j<+kQo5gӵ|i_mmMֺ鷦"{6h=jjݵYZfvj; %W!;[-X/5UBֻ4gd/yr8ެv>7_q狺sNJաY?u8Z^vZl*p$J7 >E2mpm$=A7*5LO'NMC0JYv@h"j6Ҵ\z(hB}}cRr6-@Z G5 qE e# ("ZɡP bW(`G%z?Mܴ6ɇ6jVZ0޳Nd'Wft&K^ԤX]iZI$ Ӆ\1 jltQLF BփW\ġ+컸*TQi7MK|uZ!i$c&ӴEyMhEx`:>rvq!]~7@6]40IF|IdFغ&PddZS[nE~v<&,_NzzY# 8bnJZQcf-<8+o˸ ߷\٫CF`!\9g\Rʆ+٣穹kCMFzQ&~J{TՋ͋v/4 4l~]u|7 h5E~X\fC [\?o׹7o}N|Z>cm俖lmZ>;_ z{:):e\*ZǦog粦0VeʪlLRA &4BƤ Wqp!L,ѱSA5ײuŵ|)NULwpSH#y&o2cDg<Բ1 *'gD 8},۾8E/ʸ-݊zz*j~_r)\N'lCefz\~hxLmSOon_lNh\հ'7ZLQ}~3ͅxVTlž6ԵwbngWm;7=~Y?=;(QH.Ewo;w3%z4}4ߏG1N]aq|v痓\ml4!gm:+!-$Q,>@4ɨ$xƢࣴZp@ӾΪd\L2` 77#(\82؂;g;N&\g'k_y5g}Z@3 `T[j7[~3抴x$n)thNXLnNAA[^A Q@1MbqopQ!_L><[d[["8?[ DS~ ,EO4>TlvS0g(Ω9X]*g4Յًկ4lf;[ 7 ]Go\iƘ8~9VbYN QE6cksyq\]qoD"&*ePiyy|)-[pv57s.Rp:ɬn':XJ߻ףyRFd2rwORB@CVKZKgQpф7on&cr˜}~O(P>3Æqvz_[U i-W.H`Z!'AVD5`V&s< , #› B;lS UgɑB+O."1d!%!gLE.J:thBRNEXO4܂Z .8v9ٝAbA\Cg@oFO˺&O>\6T;joVni:T#(Ajdb@Ls@ sA)0{RScc{x&U읏#ϷOA@du]8.& ,?WcX:c =[!4Q*iR~84wT<<͕EEd4 ?O<\e}48i ;|yF%|8ΓW%͛\qa8`G{/|%IiRd ̙ɝmhHz%}i3vCpV߸!٫ڐa^m|'}jf]휙_?_ǰo\x;+:ԛݢ~oRMjzz~ø76f}65*)K\o;@M5 @4x(x x4}ޅ$z0o"t1M }1iLnN-?O]_ix Bcz"T~co阚ܻ7aN9WCsH4+wN:y7](t"мUNtмn^&兀w^cqVMLTQ(oihA 6U;."D+KWl=9|ijO^mF>90Rl&ˈ/.l`lҐدt^Y":.`8Ç3)$%1y>%nsJ )hc9r`.gvzEC4y%m.qwKܫINoxkS/ޑVtq} ˤDTwqh*UUU{g&*Q]#=} KT&/0Ĕ 0X'O<9ȘJLYWΒd\NC!%p.RdYJ_f%TƬe,Knl{]r&L:gwH]=BcraqxuUڮ~^~ I[ѻb uqBwu>:d%i-cHha0hvT.FM]?ue; RK+,ThcӪ %6$cdF9?"YQRف9kY҅D deGj))a)E | { hu#Jcܙ8{,5 `=*c?F$[.$Qxəv Ks 1ggJQ`qʞaۛ82m9CYA˶fjHM =B ck͞PήmIĢMU07֙zTx6 7A<}fugdNr0OY[΄&^ONQy,g#g%K#AJ+Liz"ʈsI0А֋n9=3$`b zR0H<Mt p]&;g72vtiKPu7ϼ]Lng!,}d-47W[a&_Fz#v` hlȀ>$&ʸ!j4B$4۝YCP7"g{.xMtPFm4 flhZCDM;]dW(池vgRQ`i'&F꘷\ȥ4qN2O  wiֲFN0K]%!1d E4aML,( ʖ}|2@:i-R,x,ؙ~MLjhzDqk^p\1"&YX&EdO nJI(Mݖd\DqH0+ "j,;I&2AZ8 8Ɏ3qv#◫DC y3ǴٙmǸh{\qq뱬1itNx,7[s(p0 xx3)( ffS(q\<"+{:0 I:KnLAf=r=j <@v)4\j .% X5;Sd25 BŮ3JqGH:]p^$&v6xD7 28 FwKβʽ6%$8 S:g"U6j-#y*`9iz"T2j"Jkbj|E*nWd^dzYV\)r 9:Ƣ,ghXɴG!ܡ()e6o:cIa2g s<>xTѩd!Vxvr4#Fzǡ55cH9Mcn *0UfN %:kmF0_ h"Y 0e'6Y  HIG_d[lָ ذlYUUXSЖx'H }ZKGyUP ĕ 2)G")|8*TKAI+E[-b}=0]o>՞w 5g¾矖gJ.JXx<9ZѐZmE}n~JٱNK]!]?Ӭuoڼmcmz/+Ĉ#8 餮$Y `8(ɋS7Xܳ4xIڑ@8?ǡD}] ֻI,E]b/oM;N3 ǽ0f.)0ZP E" ԛ~\JΞ//~K?ΆͪY\1ti,nzC/n|FϯNSxW-&Oh"ÕSǻ%ay{F?׿(b@-5v0Je[6.=zM/%Nh ĺ]5T6ZN?tll3I J19st%h`%23k1_ߪZzj' #1Q3 Z$I3 CBCTD YȎP#CO+dG4 ^)7E@@+]nU]~%uj)+r|_iApL/Y$h]Ka F4y%UZ5%1E_6J{ >HZ)k&=RxGÖe`uyr뻾er=zN!|@ \ &CୣY#16tAoBʒքjQz7ZѽW]~:7O5>BWƼq7 K5)wr \X/¼KO_Vl4ߣאYW+כ9:/Xz>e[-c-V p9zZ:<^}C[ |9OU ̮ʪ*$jO[[[W1why[rJ9^(%3dMx#ANdi2 6:,0Zx+`yeŒL@3D$Ra^clz)t>!@X<0-qdzɀ6jQ!VynbΣT SYv_AJ٨u1#K JLŶ3 d";CZsq|R8l\3z/iT@6QJ=;/*ׁyzRuQ=AGhΤ &cRK*+k02fSJ\&/!ko/)uVU ) .:ef:*CFrMsjӡ)u?gG՘8|r{+W1Ӄ+5 ex ` n^;'cQkL|6c9j. AŜ rUF`6[: {G ]jSjQj(9*9Ĝ6Ƨ[ u 4!ҡ^XpipAFIɥU#)E\X qYd.s.{gkn6FΖjl/zҨ[ 2{K׳UxxIF.a'չ+@hsץw8>"27Ctj]QKjzs:[:]Po~uԲjw^vķyf~ _tp ֝zKѥ:>/vwrr7o-Xs2 j4V?xh-6o6[%%J6 ܣ~XhIIERIV@2]2 ̚|r/Vr@tȪ(2^ hnDvUg/ 102,9\X BgZ8ɡL#Ӂs;GϬ/_"FX"F Vp>cU|(2'џfSW6Ѣ&'h2f+P f%J7 >"AGo߬I[A70Yf6&&@ۙ0L;8 -:ƔMt8]'Q!碒 2:rȖQv-v]nwUD&8QB.)Q̴)Y԰Weg,sG5[oFcx Q;Ag1IO-cV#X΍WISɁ~k:_HS܎majajZ_yV˃8CݢcڷnENfdz'D)v^IFsVIØԁ`QDdZʃ$j 5ٱaf(*c1$X*1qf"Ш94B m'FN&2Q)0c#A}Yٰ5FΖr־ae60*]nKł 30>li&g'C$'8MmBWЙU'M6yzzi 1Kٲ裓Ѫ$cf|dR teUQ)X*htP1&gI!KRI$A3>(mBRATU'H4Rr>Y:FkFqrcz?>ܮ }`8m`)jh-NrR3Lm|JlXky֥5}1(}@UۜlޢX0S b:]Ï|]qau~2`K2N2Z?&,_p"* **V9ńw@vG9ll S{< :\DJ2:!fBm%"ER&M"RbƇl왜& .q#]$=kr~\YnP|8~n6&!*,U%A>aW")EVGG, > f,Sbpl867%Hf1|a.ұ)sx/5P|\|v*Gd]Ztz["U6Q[:uO`'ڮ+P K[mW4{vB)#B0+"<uU.TN]=Eu%ЕԌ9uE䒙(ꊨ]]*Av +ե5z Gbi-X".&Qr4'f֩Cw5޷WzSͲ~)?R]Sbs"i?qpBޫ3LRKBX^Gy(Ӄo&4]]ںJд3Iڃ<;~vfP(-OJϔ.4̛˳z&_-2wlUyԤ L9au'SV!BRojϧWWiνVVB-D=_^-1y)eL_ԗN3'6~.]#Pʶd*.qks:T3ߍ;6= IM/P[y|` -kQ~yyJh1|}1<9Z=hۇp&zJc_ç50=><xX{XQHw.н'cdl8W].ųLhUOh=(ȵǶ+_ɂ84ltZC}o} uX=#~ۿm2dT *.鐋_ ԇe,3:'YP__>2=m> U-U\[Z;uiBi;]4ܐJpm3k3@Kz7;y*4+]׶sShL(U_̖EoU`=g*G2{8\V8Z881]2[|ܐ8ftH] -O^WBf]tbwp0] m&Zo+\?coٸft%[ѕ@iv欫gѕU(ؐL] 6JhB9 g]= !4h]9JhMѬCUPqrA s|阬cOd7>}6!4ptԌ=itsPydj) Q JhySvBx*RRt%1f++\ۼnU-o?;~u5?ڱIW(yb;]=i7+@ QVt%O]WBiìԕPPCsҕr3@S(ymunHW}3ȭ xJg]8ޟc+Z;+tqʲjk0(]אjEWB;+~ʱ }iu%+ve9:H]'-Ȟ(}5Von̩2R7OeW16szzqSCZ012Fz2gC)e\?lTm-0fM+S /J7g>ĩ`nplGWuܸZ3y] gu,hgJpYt,~rj̺zEouAqaGF'Pњ]#]v쬫=h5+;ft%]+RYW+fӐl>>4+ufPWY[؛ft%+5y] %!]xk pTJp(Z?]WBue=Et%]!vU*+ܱu%fj?D]9ou ۹3(+i(YYW@qmrrIQX(sbKZzۧX5ij'\ XhmY(L`y(ۉ*Vt%NM]W!hh !k*ppJpC3`FQ4]-58FҾW5{Z;N<)7BWnc"=oT]8Zӌ0rYhO]WBiJGjHW6Nt%Zѕk58`f]L ۆt`cl3\Lt%Z1rr 8f]=8ZvguOTMo2@k^^%/ W(CeW/7ŕe!7/_xq7?QNwW96[.{㯇_f|.w[:k%; @o?.^tw2on$ۣh^<%Iam$|Q~"4x5Ԣգ쨪HK Wh'4 Ri{[sy(45Gb1{#)NjUEx9W+n[{oz}̶sVg jv w[uzꡚPz7j91tUϰͮt%0J(y?H]C&h^CKXm81X @ >^e[*L{}h hZhB(ìCԴD8.~LNi^- 4=6Ń G&9rTSz>{m&)NMkKlQKy&n`TKKf؂bo;?Su`] iƍNFP(CuoYX1^=Í(c{8Z(v#tg]=bԐٷ١HqJh+<uJcl#Њ֚J(uu2cҕvtH܊u%v XG}C`׎7] mS(Z=pteUqhHWT{_qnAڳEWN}m)3] .Vt%nARWgw8é(F1:FMÕOnnobp-s+M ӭhڳRi!j:Y?lpiEWBu%q#zl#TCsvhu%/HWaˢLj(`8}or9ine&a¬=YbQ芬Vַ+j㧮+PHPWZ]Cgߌq+`+<uelmGWU+ڨ+P-u,b]S Vt%N]WB㬫ԕuJp]3sW5{J(uurѕf 0u] ճQW>O-4FGVכWond2*f0~.|vNrtȞwΟqS]ohBwoΚ>cA~x5ևyoU{xUkRnĽK"x_RZWH* *{vrBy -g̻#eC;fŁWkG2{O=#  ]w_ߤd\ H7uY/^vW_X--כI[wcMCd)Rљ\ʕq9ۻu O!x;vB}^W7܏kyn՗%^^=.'DRzW"gn*d.<g-%e l7Tl z燒}9eP*1{BRQ(㔶s]17^~ ʎ¦̖HKA{R^TQ:R,",0ؤS2Lkc ]ZS:Z Jҡ0"1WHF'J}NѤhY ؊]@H.]8[E< d&#s8U&Q {*u),3j3bv."oLtߗ;T rJ фʌwSkِd*,BiwІr^I{ I'C* ʚT>OlPZK2GC-> QE~,u$CHw }C1gX2d]șcW4|}>U܈@Ҫ:枝.K<*.QAiĠz >RNZQ}§rmja>"u'QSv)$qt5:-$Hߩ6bhivVC QkU {ė;[]FO oE{"GfbgVŧ@ꕫʧMpԧ;Rh)3jrlU2ii9Kz.DwT/Kԣߎ0Le*yED;Vl ]dr3OYBaE#UN!B; 93us樌i5D9q"C]."Y bP3ǒPѶze&X4!dݩ $B2ZCl(e]< ?s]*ljQ7zԍztJa)}JVȢ DѺC B/=zjAׂTN}]  )w #-e*FM%RĊ;Iu0*0LF8p%S%,d@Aq.]}U7w]FT( XV8NL6’  mDoj,vEc mT}3B5Dow2d\cab`H5$4(fƞ;fiQF?BqkCxA7¢*98jՠ"{zbEb`|LǕ}E0B@tf9pq>aQmx+r=XНJ-*\Ba*3[Q.)˲U&,+yGRt!}aXG])g H6v]{6Zt4BPA(}]qI;=A5S˙ٻ޶,W`mӮC@cv3 bLz1DܒijS,ˉX1ERn{.YH o3M:kPX1sɺ6`-A+k| i>jG$6|45Z]ݙY ԵEnBf=0S#;5,FZ\)8)  j31e8оs2÷A] <%2`:39]PnxВCqQD[I,Bј2!SAB # RT",Az|PPʭ`ըj1+ ?zyV$bʅs" Z)ozل(V1Zة !jYPcx*QGb'Zg=ku @2!s6,ѳdD).XZ#e̍LmWjF׬U gPqQ=byPbҧ D2)I1@2˷o5(V8/%Po5Be_́;J*C5*mP zx;XAa^'XrdFiC/ڀB\D/ĺčt|k*^kO^t(TuJ]1Jn|pQi0iט 9$UUR+DL0IJK1!;FfAj>#a'T$"vB%q9@B؄r^ZQDS F8Ø|`AAZUB?۵q:],.WڶҠ?RSh<6AqHZ%g(N߽zq&zӍGhxi,[ӧOum>N1x>YJ44pZx\Wt(b:};X6! fd1Z\RO~л|3],9yÇh TQ rU"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN rH"FD@YE ֹ8@?@*& tN "'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r N>98 RTONSt9iON r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'8\Zoz+~<{5KMkesqjAݩ<_djd\\R?ƥxoK@;2.q鳡^c>2dR􅮪:VGOW@62Еb> UC hCNW%IҕfI# ]UOVcrDWCW/d0Ԃyv+e_誢5GҒ:Iiv?WkkHm~«Vˆ;!]e G;jP!do/? fr׌cpV膋Ciڜ7Gs-W5c WtE^UVn1"Mׅ+OX1pEoXjutUQR|tOwD=uu($ЕbzΘ`jG 8vt#Zy>bPr_oz.7JW ]U*Zk*JOtut%4 :Ύ*JNVUWzCW]6荺hUE% ҕ#}rnBW2NWbJ+iIӏܻևLM0$!@y00 &(vsnAGd:_f[(mA'~^gOjF9TcT]__r>]zci׻1iݔwe1CӚW'gЂ|ق_-+vדT4]^nfa+gQ=?>~5RبMCnj$OMq{9|:.7F`lj8jҾy>1d}R>SqjDE*U*MЁ71-:r}07L(붨p6*1D}ӊQ%-I,QyfYV#FFIojs";_r47?!K V pͲU4j.-Υղv׳,hUE)ҕ7L#,n_tUި+}MW7z~ [vl,f76Vn(=.;'֡ Zv̝ ] Wh9v(!:A+zDWl{DWLޝ ~{R8Itut%VO  zvtk+e{DW\|_誢G_ VS+;;;?tK*Z)*ʍ :2N -{DW,eլ/tUѺ+T]&]??2*o9Tb(\v 22H7z?ԙŮ.smrc>] ]eX{^~8^n(w_?>T/|{˸u:_ONs{//@fn͗'8a]]n ~Eg{~Z|odi.^}WmwY|a\M:jZpe,I&o^==vV?[ѐ}񳎍'v'+!2z٢߯u9?yiry>M{ut_ Cqy ぷ20eN*d#]\k8UvJ#M)n+G܋Kێ&9&/͓|yQ6hiz5]ތe0Vσe2X`j̕k(l0{4:9)CW$Ӓ&a#j?Z°KmLD>_}!`,۲ܰ1=WOW֯݌wuËX}CKʘM]JN%tӂ/фQ0\;x1M|m >5ۮ"mo gk(Z٣R)N>x>AOx)䳒UMat>\ -g)0{PNI2oTdߌy76)4[;ރj3odcXW"5ۤ"T=lZNnT/A:߮:$ݡ|[|lLPR(-˾(oQֺ\br3M%f 2Z c"ٻ:ϷA̝K_9t_:]QHYsL̙ZlL]#V|*ZJ冻Y\-]Hi읏oj/T/iz[R, (ϴ6b6c/E]{ƅ9|U˽Xs#&{6Vj>٣xp+`:vSĐ^=^u%{׫fJ.^Ma鸹IiT->,8_ FEtnڒ*,9 #d?n!+ܱ o$d l2 rM$J9"IjKN*RgJ&H"2K9댰1%W!+2jLJX[UZ]%2[/ވD N23!Y ׅlֳ"M?^]moG7ë)~Xdu1OR?P~/q_:NSrEi2+oYR9j&)Ej„X,“ETp^'11 ϡ>I FUI6($O_͡L:;64e;z]]o#;r+~],ŏy 6O$yE@HHSlKۣd3}q-6uxNx9hǏN#Tolo0enbMm@>,D=LLkMxGQaD⻩?beX{_1z 6wTd!'ڙk\!w?g͝ݔ.Dw./{.ox6g`S܎>1:v-|wN+3bw1*mؒCR(*^F )!֝BPd(ۑ瑏~`0mv򸼍i}t+OTgbZc]t0u+; ;PʰX9y_c^#=刂=e5 S2e{LPNH&2J'C,Hgb@|YQB̐CSaMY5T^6%>f^J+6VGmVug?:=G:_n |@%Ԏް"SG7Z~U~4}&E6OS43+10g7*X9tck;<CR9S_`\ 9Ụy<.ヘ:cdN).?[XӀKCPz"QP6 S^liEX?%dCZ(Klu|pZ`dzAѨfF9ϟRQ`Q3\x}񅠤LVAg`bAWNLk(u]؇3^r9KjJV-sxtTNLB'r0Td_;|A_?fa{> QĎވ9f.V 2؀SUt֢2Beң4Q3՘őטգ ;(JsEPT1Aeʰ1 X+#dĶ?R) IIXg%_4 "x-HkPuUf'=&zxҿpצ^zjӢ@va*S1Q7^XH;~v料Nxp/G^_y.~d}=W}@#:,;WwHF`UE괧aM10v&kd 3\)G!Z,K KAtF$&RZ4N%KVml&, )jPRуXy[@Pb*x&4ͺs`=*p|c[o?rj*!e,TgQ  $gsDaK*%h2)L.G:a|۵NJK o"TZkzoėNl$V*Ti0)Z+\j٪;]-%.V,mX|U0[qrT0SVrz0C{/wݮ}CZaG]5w^ M\/7C{K!?+A*I #LI(CԌI;&:࣒M8[JaIQQ zrY*>J \):L CajͺqfXld1 ϊ=\@>~o˴Ol6:y}_UΧZI[2c,Bt2"`5DdHʬcb/*FdMel2:MVzLpH\X$+I̱vq,jƨm'Ԟs:%&)*23 M0M^d)46,IBffUtR2P,]2LkxM!RAEuuZ^o6N8Dl6>ED݄"KA(/7JѸy;pJ+QNJ2 Re RQ6ΨYcAղTβɆIv) JmU<}Vُ_3ȸ5VX\qO8L,ɠT $c)Y"g@bT r2.8] pqcfH<-{1j}[~GG@)-!ZZ #D FaM$H N[M7)CG PоNVv7q_F|W^}3<}̫_r1HNjVѳFDxfaDVGqyQ9ӓM82>fxYc.KH[E]:q 4D0 3Z"Yߩ_Q]j~wW] u7g' nҷ?/bbys渉OFwy2jmp'!mZw{)#r0_8زY2*U9Ϸ7+w`L܅|Gk =݊zw|rOO9ߝnڡgҺҦDZx CJުzF¯/7ֽHNhLc=6'9Pe'MfޖD4J8J$%d'dKD$q4q/G)HFe vTE5HZRYzQْ)AP(%Lm@Te^}KfcoAՖ\wxx&jeISA] : CZZeHo})[?,~ c:*C09.*:BZ-{)czK o>s|gd ҄Ve2y4FZk(\26 I)&8&1qQ̘T:B};jX-?k3+:i}ȆBL:GXcFZe ^bRMQmw~e>st@<>746g^ීy#}($M߰E,糼Q6mKGJèuI}wEMPXr..Nwĺ׆6I"tߎ(sSzq]uLyvzIP-Q lki{s)uq84lXaؗ5<yM: ; Y&=BYul׮b>_zܣMmQ[AzJPfVFHP[nсڛhK.H,ADQ(jt9٢9XFp0Ł sC| f廭>]ex*} :{hְ#"!yX؀s~)PZ:0u5nZp" " 2D&24CB[bl9̂Q|Ao<#E dt9)9rjYC\AD3*,唤Q+Uʻ\UT؀@yDP<0cxB3Pc c(UR|gb4qV 8uH] u@8XPlN=ɃcR%0S( ?2h,ѴDJI+sA,MV򜢌d"VqXCմHNMXY-{ FCE\8!!9"(uzZ+Tz+buO;ְю0~V>?/hW4y>U;yUo=8ؿ8;_dFKPC +݅E3)maq1!~D:Aj+]aWfu]ԷeX-R_yS]dѢп$Rs0g`NN.ܼ/fuY0hEytb=C $KbXs15CU1oEpT}l̿ YCO .Q2rwJ6XWS{^$ >}~5(yqŨ^?oT;^*qLZl0/~p~/~x{՛7߽~{L>>oC6 Eyw/[0| яk*˛5_&ZO|ryUn(,qe( \WŏW>8_fG%*00<0aHQީk &b9* JSGWD2 v*B5m?=J|Qe8=\~ ?{ytjڅi)}'ddž\ͧIy'JQ@ qobIQo%c GocwFtu^z \v*y1퍋+}ft2r妡ru2:}Ɩ7!Q[AlPeօHVJS0kPF:x=/cZ[Y͞?wqb"A )ql;mvi}`L7D<;1v7Zکamo#Mv(|g1F_y)ax'󃬵ފC(?pNHf+YR؄:ݻ*iT}[Wc[HCpAb"&T!tOZ#U(Y]ݎwð =0?=v5'Ҵ4qn㚧Kq'=ej߇{#up'捝L9l4`EO#1/q;=b hgv(΋ ;"U)(a8[_d8))xp:'-XA JsHuK\c%IIO""H\Z;n]I9[ª7VKt'.'bצ쐭E˭͟6gZgڏWuup"Xu4uiD%KFh\G=\{}+V]mO._ f9AvPE̡ʠ̡V]?TQPsW_E}k<Ι>#J!y)RVT%΀ M+=.BNHӊQŮfeRu{1F{s*Klz[1|5ƦPx1A{@/`[U28'8H>pH)?[Ï{i8# ʋiZ^o,=C@ǹ)}RN(VRd'xϋbqp~ǂfS{+1%6̯߿:ܓ,ȁpPBXZaiUboճоggu&(AsebI(Pz0bNӳq+v0e"ʑB*s gsSQoK\tFΘ.']13g(}gib&@~"]evdd*=]}=tWz0o`X'֮Bkm`jvt{ڴU,CWTa+tJu(<#{z>tńfv2`-;CWUF ;OW9k]!`BUF̮UF 1 dN!BU+hW*| QjJ{zt%%BRt2\hkW"=]=C¹X0W13܇& cNW%𞮞#]iMܘv*-)<`fe'PWb/5yܟ8ޱW,DQ>6&IR>o f_ 1 mPKe0GmA~JeKlݖpAV].宅ˣ9ѼCt5P!:CWxep{(̡NWrCA=} ԦzpባV>Q$Pt~t/ Vtju2\I]hmQZ%XMJч,=&'1v</Vr:;~LRu.$?ƏU>*LJ4/']:MxRDSʈqU}ۢR[@UA0}{0MWtvEg_{H"{Tz-d݌W^ Wu/S|N9n&R(`brQJttLX(IN\j(q;A=f]kH۽vL6h́[V:u#lI}P3͙i3ԝXSo.=7Bzt$籇=O!6m ?[{s\?5;vƷDj=#okYm6sW!"*|>]=$ybkʔZB%{ݟ/lU>{t]Ǒm)n< 佉Rbތxp&sQ] s ǡd[LK%5bt~&r 9&CRiV"Vu!芊T0qm9ܗg=KiInQn> ?|8u3w $_C RRl$ͤ3Ran@h@?k GW`hsVmx:kU0ΔJu'1V/̓me8nLR"%Z6,7%Q)x016nƥʦRloi~0m\69_hgjrXK-ʎŤP~dCb˹%WKnXl̜pbLjS:ӂ.IQW@mG* sDУE|MrmllQs]5l![c3h|WH_],b L%*fnoʔ"$DNcrq>2YuQx9WuUݛzO{KW<{cɩ[RP 3y]\oJj>OKJ]?bj9b“R"y:_FdmdL&;5ȸ `im3W>yxuKity!\C~/u<2[0ĄejTb(x;.7utl@Qo와dd%5`qsu.Hiڄ)R( e+FHE/Q(-VzlK UC5$f66EEklU߈oP:"Bo:navz~ߺhQ=%.othzl[(]β񖙻w `"=Walu9/\bZ=f;xN6VRfLL+4]SzwdWt֤rFacJS9Jl 4T+q2&9]dmdVOU,fo;*4J4c\T崩[65nUxB#j|x(B68; &!sBA4j(04t-\-wN =K&b &&[1%`CA"C,}tޅd*rAE4A/bRcL5$0#jWorl8ʮJQ/Nv)~},5HdM(kt!ltA: c+8H|LkG}~횁9JilKpɾ$4TsbW*a ś"3lI3Oz"Wݕ^דZC"[mŪ&Е#"qMA8etDHZ/wjw*wp\Uy-մV #Cie`{RI"F5GwǦbc y=oMfk#}ZԩGD/cߦa[ €77,Y>J2\x[~<ƙ iy 1LCkLCm~ZUzɴo HxM@Jq^q}bڛ>0)JdB<QEnjM)M'] z7gqu5lB5!aSsRDñ:1OJ Z@qF,ۤ0VcbƷU}xcO)o2\N=奡\sq2WQ voZ {| f\5rz$u6r#L62A*2e$T%%Ru٣sDnZV{Gza"A@^[u!vTXƱY#CL6Zz.Vp))6[qN>'УyaX(g>Dc( ҲTlh4t&`be4UZ< UN5;c!)a{x&&gR`)iK]4mW[ FGOL{hHF瀸jXag>=ϫ2LOr<=:fyϯ_W}CDik7}ۙo7wɫU)&PcH_,7_+Ćzsc" + h -bBDoBԯk,@&ng [)f]4o*AR,-+j r%೵3*|0x/ }a}]}xKc_>~__0峳?N_Q {l17$͂_"&WQPIl08Yj%= D_C'{8㙝wm%(ѥ4yC9HRhRd}1J/-y,^{vG ~Ӽ7\%~燪Sڸ*n~Вئ{2%8Q }/td#MkQ <4j$<ɅbRm&拲45ؒ)џR~_.N?cvMU R55ۗ<[N\Zڢm-R%, T5 1-2)T,- nj6m2Մ!߰r1fu%ΡL}Ȍ>&~ ~g}v=oƒ _\K{? wi^{+xm./ںȒ+IӢ.IEXNdNM.wfwgkgg&h_7h:jhOGd2 FQ3dI+dI61bK[:A}{ؔ-B#Zh RӃ3X93.@'L ٯ '0L^}m"?~qa k s$pDp.*RrS{) jQI({ËlKw` x5ۭ?z;HG\uQ'c 1?Ml fPX qL @9R8,EkM5(Ɯ8R`3*a'!?0ω1(O-4EA *!4 @-((e&+0 FPE^"^s@{4DcwK9 QR;Auot(_b' 7-Zbᓸ裂ɻ类ս%-ww{;F6]xFg<<`9 tcyD`Q\^xV1;"J ,90~x2壕8ϧy82 * *I4xU]|:vogX1&XiQ #t(l,1V`7ZbB!PAZhS0jO0j5\"{hP^[з%{~< I]< 'qC f1Q3c$+C#8b7)8*V_eՅ az:g<ٗh$+&l:8Ui;QWQt'WoO0F Fxp^b\a8X#G?~#|;@2V}3 wfHnnNM8 v62.,]`!L]`( Oc;һTz'H~_a[:v|R!ц":LDm֯[Uݖqlټo)xzґ֤{啺q)o X$eyZ*xxM9\&BW`6LTPsͰ5U]ݽC>oHkX{="T51JU:!\U&;2;6zuZ*ӼFsuh>TŽi_ExH-8xC6T3WbVGfÑ10'kYY9PHipϘm`N;DN+P^^=џ^NOGQ=gr |%qΆ@45 ck-) s8E ty> }Izm<{ϫ9_gPy}nF=yԜR)Y﷏W.ʧɟn#3*gRbrAs6rˣ]+V)TYWP4D$bx J 'Ns!Te BS0m-Q0^;hUK7ԟV 9s^xڭEˏiM??;p`J(Ɣ gVxk62mU* Z Ct?> fL5t+Şi6xΛϚngVjɬYo8"hweL7t:_9:[:2KιJX^"[uH&ԇf`k&B!gh՞{(@'*WL#,8ZIZVEq7C\Z^_.BHbmȆ(|""=C8n~ x|l[rm^ގG-0V_o`5uȶa=u{kQjOqG݇TDn5"sz>9I7<%ܡU:I;z ]իhtϼLFǬSNSBx)..mvfri&7oi&ܟ_Ү|fx3j7rcM4e|Ie Œu'˫棜h%4uj o]&XkI ?1.5ɛN|c o1Uytsz)!aK`T eY@_gXw6~% 3O0BmjjHm`H;MZ`22FD0}zǣz llJz ۠U֞CFkxSf́|O\l\dxq6.uӲUKhU`T_͋gϿ{|_ݛ/^Dyͳ7/{ Ypt5 ?o[ p ?e_^ *v7hP& \[.oP'Ci"8m J_g?\t.|J$oz"2WpxןdeZh`Xq`` Pȝ2]0J.TUB%ۈ1]<0X3ȐOxdPYkhݜ{[xo> ci_022ΣY*f8%-\1ժ G1ccc{-j>z..(KL-cV',u{Ec A`P/C{u7-ymzDWY'ӭAKqA+ԪChJŢ$vv0Z>Y?J~W9i w8q.mU8qLOa]E2+[%Nbf1]vy1| i̾UqoEQ2}r(+Y _<{AvhwIv"`5~fVg΁w={}V>d090c <M?Nt sBgn퇻XE*p8,_rP++PN)1Z"MFc2oQym*^`Z]%v]q'ARz)o՝x 1/L<57@ Onϗ:^^7Sb|QX(C0NցJhX9PK*[O3ex;~vtret[0u].b{m=8pX=p^+3v .iV){{[8HhJawI&8ڼ "kY"My Ւrr.`9&WâP(?jϔix.QuUTc7HeZz)55N_wOX,lRy*і%h.,/eUlZE+Y(@by*@(@I$nOoPs!ӭ(BqdЛBT(D NV Q,+Do tx%J]O2!6׸ڥ=`*}/^?|W^pW}o)]Y5w ~K*&ۛrj<=lIUVﳣRxlrv#EsrUF`Dl:"aӀ./bl:bIсMlZ\ MK 6I(-{65#o=,Όb.HvM(ե(<qIf\ AP+a[Ѩq;o"`0tk BA*[[w@WOtSbO`BuXdaw7Ot,8!YLЁd MRL%Sv,iU̗ӁHaܴobG}4&vD&6`KޛK&ghbk|DJx4 ň?v*b)7#©ή0e][]K]{+~`W.=& #bW0Ѱ+@W>y뽰U;!.sdW7ڇ2]R%Wɮd^&bEJ.nwmm~>+v})pОӽQ3dF\IvP.X,ѱ(p`Ykiioȹ˞`t0̐kApEsY3Jn=>a,y-e6ip'?O6wpzViNYQ$WR+{&lR :e0sC&XQ -}/}m~_&AX 年F7ٗz>Iӿ-2?Hgsq1P,*|n\K}KB*թJXESE4`84YfFH٣)caݤ{OͰ{ }RLu8]t4Λ=!efy9Ljj5VSNU=_yٹ wbLrg*xT;mH4VG!ۿFAQ|f2XZ_Ss?xDpp׬դ1.pJEkG\3 FqHSp$\ѬÄ:L]qjٰ rmR T hP* uIRUs=nlYgv*3iۧ),{m;Ui|{*]β_7';?4 u&CLmNxg%Ոo.J?ozntWdkPT r % ljQ;1/U1p àcꍶbHF;Hba2&Dwa#f]2KV s5NY/ޜm㵓(>pE3L/^>}-l*Y[UlYVUޞVx=UW-l(~0;؁y1wM\bbz]2cw{',])hnB[][y! \˅zz0'ymVZC#okm$e^ vө n;m[BJlM*VU9JYmaJUH'߷Ϟ\ KK.cJ%TgY@Ow^<6 =&KGe_.HaQcWK*<*7lxIQMSA5[e#Kِ&JG@U% ű̥@Zv7cG;;r*z_"0G-Ym@g ((o$4>gjgId6EW:c3{T\.Fh:Y7q|aK[@KTj@\4SHJ4R D lƐ|e{4&KY1yjjT9HY#CU(&Q UX=rulL 0dS},Cb*/`!3!zǬ\66n_F:*kyMU:QYɏbޟ?̀k4:LCXuDJvųp2V}qyMfo9W:Z[ -_Pj|pi|[ԖX{h3A=72%y48y)T20<~p<5hrPjZ}?k+Y}!9e2,Oft]1bnN$Z* LqzUN&{4;0e!Ǥ;Bo:Dh䎅.1_N.d%]9n%)H8;bQ\`&B _AxHJt6Š -gR%L.Z ;S;/vg$.P1Bϭͺ/޶yڰ6}v3u/([K\2{,x:9gsE}e-8=~=f%\ :7ْUM6gqZ|E`ImZR|=W|4_}j_Cr 7lp<B^bE뭯`h4$@QhD+W;,gbP qj4FHȯ>`聴 BE;oE  FqBK`2S9b]sLeH5hRٵ4΋ٲBڽ}HdB|9>\[{jaV+P%[k \k6Bɱ$T!bWC-|k28hk WP,F⵱&-gUΆM-47jN޷Zb*,xx}LJ{Zo=ͫE'@=9лJV_Jɮpt46RQBV5= Fۺ:/ eLY 肯5glI p5z 5fͩp/`1UMW醅4cW,tIZvoW3ސٸ88xpӌF̓8aH}e&UК!R4'QK^P&Y<kpVb6&$2 KY`8OC:dTUc,]RI8#vy<G{o'y ?L*Gm!$E9DlrBbTOILW=Y>G<^}_wyTDكFb)}Rj-\!!lؖSg&{JY䢫!G` 56XiHTyb5,cR5CrOe>Մl]N/>,7;,7vKh#]YL&BVCf7"ȴrDcT - GSK( мL{ۧ8[<drU7>>bN~ۧX8ś3<0*%v,'Zf93M?]҂~9M C  (G?;#!.Ѹvt2Vi dpxNYQ$JW|6) 2K9ҡdmB`JΨ\0x)O+T.st1UTA&ʏC֋85GS,|{CWy>Om+Ym$V'@S߼P 3 g̠ +hIVa۔vt6){)CqNɽv5wQSz3-%sV65c-Eش͒)O[+_ޖڮoh54@/k+\%#-DkּM[ݤ]hu<%{ӕ=suu2/MZ!h ][ k/Vѷ4pm\b4iCEBY  D@TqhAj'G ;[.H#P+46<7J^/An\ػЇh.۽d =k oUOClPno)7Z)<'- kc{JBJ0YAǵiUskr -OSa%Aq)H~Lﻑ03 ?>v}o΃io49+>Ceܼ֊}7:;r9O(7x? p1M8$lXf#\=0pн?91LC%a v`Wa1W|FTp܍k%8àn1#u T0+`QE9 ۃ-sSD H{e!^ebMF<)^qgΝ9?V+0ᦣ.vKFpC% X\2"ys.{l?(?>V(ze,EeuPm H$4JSNJmA9`qh{.rn#,ۑEw9lEbLRHDcGٙ\wɛewfƑi}lOdG G#!`@/qA[T+EǸ.O'6!2`K.-ʠzs;oJVnU0,ww@i4 ;$*8aB!ghD1b"X LNJ[^?_FM(u_?bfC10,!]Zӵ2*^Q \}84} 6\׸]jd>Eˆg[,Md+NI@!vZL˕ P\ZU6z4Ͷ| M'24wnܩzgYעc)))Gd=. }-+9 XtYnDFqzY)0I:,XUi\^ROsaR /:|Q6sLu;"E kN*"+L̰4 ؆ (O#@2y+~'x p}(wanŢ :5( ,>.ޤp['1.X +3/MV'Qeb&8K6)F%ϵbBhp4Dc)"23NhD; ہt{a;EGg3,C[X\82< nF8fؚ=VT jcSRЊ :+h%4im  =&$Z_/`:)LKtE"QHujRumV?OU8?m8\^C˜ѹ0gKӊ S}4gZrhX OOp޸0<`$TkI\Y[1MZbQX̆JG6pp9kY|]l-*Vjk]5 yӴLsP&.eHFH;xz?ßgZ_jTN|3DžOAo߼{~o?;cL?߿f@ NZE{ =VEѲYTE+u۔KrKWFhi"/4}f.\ήWYЏG}0`b/앆FkY*&0XPjoM*JI.4T4B17L*4ꮕ/M)G9/~t 7o>N(Rb.Zq],/0 hI,_J<}G)XDۊ-kb3_ 鼯\,9!C_UGer3-}de(e'KU0p%z bR U$.Ą0޵oe5zŎڀ!μ$^ ʄp2FJ9kwI-Q iW4?áNvȦ3 %<;azDt+0J.*1aV:eo&_aрmP%㠴R`D$| 0O !%^Fd9q L tȏ)"Ҡnͤsdcw5wH9lO v"EǾ7%>"S:0YQ)AϬ6 myTH<&< BsLnocI؜TP-TIEOٝV<Csb/rEXd%R3RF/ST6كL))eGR('MDV4 #Ʊ%HWg$ 1̡hf$,!ŒSP`B<UXp*;c-L,,`2b=6MVHKDaglySͻd#ޕ}x>x[N'pKf ^5tk--2̩PUEWxH[lfݐu oh|^i&ꊐkEܧPmc aBRlFw.,UJ]֩; r]ܭּ^kc.gyHEYK)d4ZhomzX֯n؛pV\YR19,鶀HFym$TS@5U_ܲoD93"^q9cφx9+s!^Nj 4/@ePOP,M-wDN[ZC.G^R:C'O;Gwh#f򔖽An"f2R%IGtp(#XfˎώC5iㅑJmc}J!4E N<< v&}<R3,q$0U#\DZJÕ62/Ɩy+gKudF y#E*)K,^+CL 3PYey@":X1.Q [=0> Ȥs0U4.]߱*"DHE "JFc&=jǔjWS^j]̡' UF(`C)KA hHm ՝cIg > ͛AN/nN9Bt~ov/^.vLJXlɌ%8߱LmK3OcUo=9x9DƭBdݟ NF@`0ٹxK %mq,Қ9yҰiiNn- i&$v(EBƒEmxLg>&oGEI ep0g\KEQEU`)Jdi&6FI׫ПCkVf@] OOܽ4q_}2n^DgYƜeYƜeYƜnj9˘9˘ɘ9˘9c-c2ElIƜь9˘9˘9˘li=671~6DI\B 7Q]&JRr^ 7Q yf0Cd3D6Cd3D6Cd3Dexj#r%Gx^r%Gx^r%Gx^r%Gxy:\/KB/;蔋bu5qYI8\oUׇhw?Na\A^b!I=hx`]||GrL2%3"IR*II9SoOt!\s:!_ 7opǰ@T!wG#KUȇ AlsBjH+W=$TSVcx!J3쩙~zQB.0̔P)9N;Dee nB.^e)DP^$e>oBhr='O%\$Zl_>(>20T _$OH8܎/L_L6LZ2H+sfLCfLRQD Ky"[rB5:Ypog2Ō )HZ}joGL 2hs)C`J: J N4/z#7s譭ފv5K㺄4̏ȾoyltlK!A|pJiΉ9]Qr~ͣy܂cjuM-^ܨ;el4 Bb6 )6jh@Xcd8dPK]J_oA 7W+@muOO"~OSi^R`RO$0gj?qz-՟(SRVg-@#Lx1/f7T\3Уe}Ƃ :6\ =;P-}W#ʯGQ !|Rf&Є!hm|Wpeuߙ1Nzm ]4L 6;a#7&DS#`7ɪ̽`)KH/E%sjlXcgrf|s/`ϖ]Xϙ *g֏oQ0Kڽ bzzBZiXLIf f:rh@h~6UrOrx)&!(ydd=n +̱ٽ؁^AkHIF*L-ɢqHJ)J&YhJ~Fp.@xɣH@h )Kn|L%X[ J8Lt.3[VgǢ^3Ot:|9"@"[4ۀ\{Ճy-%Zgo1SJUM&"XD")EVE7,X|Ar<1SRpS2א%H1a9-!I ϵ))Bnķkզo:60e^vI[b)M#g.t4EUzu7s⮖ET6=+W#נÓ f c0/CaP~kߑ&*'h֍M̦JL[mos/F) }{Zb;A$$$B(HrT$Vkt!J5`BxvR:= [c4oE,D{ s`9!~2ã[Dlu~̎Lj5Sl1AjdQEdҴ ccFex&؃N{Hn>eKY,VY.-@zK\5 XĨA&'XfZ ս
  • u!ݻ{_w;]xot<4`);\4Kp|,jǢM.WŮqn57= [pu|WB3wo@MߕM<kP 9Otϓ%ҕs~~ߔʳR.wJߊϫjfE <crܪρ)]IOL^ $pTae[jE-;/MݨUo6Bn֪g;1F}!N+ѳnLw9#4y_!vP^-NyuZGrN/ [Q+\V nitV-3e2=e"/%T׿MVq?st#7cMZaߓi7Lfx 6L>%6݌htDsi\^^4K':?|U=R7mg76|p>'Ow>qeO%9!=;C{?;NƑnb<?2n`a8mYKOg$LFߧO0?r/HyF˧cA9;\ 7 C~weeig~\P|JJJSs{Z]ή|U_jWܻyOo{;^W*£U\HB_ㅊb`I:X?4h< ŢF\&XgLɸ+@0䢯Vb =X~i':d Y3f'<xFu X44Ewj^k퐎1ɎʉϜB# id\ m{v:':B j[y qIZ)NxK>)1/f6 6$X Xc H =c{u x/\<bܰDCi"xR2(kNZDrR13jgҙ@_JD.Du) ʐYxKU4VE5qY}Llg>BdMr)E%[iv;UyBFp=v` Pf06d~ ъA eˆgyx\̓iV!I1`{!EqR+Zب 4Ifl`XCDMZl~2Xv5k^[^G֍"&F꘷C8!H `H#GYJ 3UaF]2!CĢm#Eȗ] ffoػ633QEVZےrd~Z/lɖrr@bu,>$,L H!ñī F >FSE1vE#V]5]{xg2 aXPSYz&%Dr$wI%* <פе!P7,I**I+X.*IΒIH2X@sFl+dHzq27%_g5.U/zz׋wV6igQ)sHsH (*̅׋ubWa5Q VWΞ`}*"jpy7p/{8MhH'lv.px+Q&-46.QDtM:&; NjL LˌɊ*S.UcWW#g4ȋ9ҥ7(U:YZga׽Mxy[Y*aՏʹcq$mN| ,L Y&0P\yqӡ,F.{fCl%Uwn<ð[JVfy IY$Oln\ZF|/[m;䶽Y?-d~|Js#3^] !Ͽ/ڢ鏏KvVàt]iv`z{q镟 ?|<1Ε\s>GtxtX"^ރx4m::zwF{FocF2nHJl0b0^laާ2z4ny|1Qp<:2ѓ1g˻&Q{$7qD{185o (.yP<5GPx!uhzycbǿ?8r mz HH}4 |^>[Cyж4|hC;jz·W֜qoglI@i- )K?9/֛^imhX6G8WlidGJ%F4L\7{̺ LF_skG1pvy[yu:p:듼ral8'\o[|t8IXf2F?_yWd#twXd]+n-WgN|(bo.%!G>R@N4g\Ri]`] W8T%8ԩxb36$;v*kAƌ.@k̪>yJy'R.6] dT!FĠe$kJ'[HYC @C-3owQ[_J naѠ#\!CGf47e-in ̕fd\'Ɯ( Yf&!0Q)RΨL|e3R::\Rɢu>x3&y:8DWt cvKW#&tSXX"$"gȬuG! OAD%.^XBX.4ҧ=tg,*昴QHRqJPJ Z #2T&'ozw A~>^jo̎j 5 q]c⎻=;1gTzZ/z/-ZU[*d xo+TCt5çL<6רdLX-y}}5-v |2Уb Y"LaBp@Y6FXk]BA`"ۺߐ\ȐDknA}Q)3('5 |[IXHi)"FUB"^!ċq}75Ay-+?Z(PeT4u׮3'W)hH>OJ ;ɹLBB}ҿŹdB'&텷K]2fiq-rWEmtĬb[my&w皋'7$i#YYT}s'9I\d]6QGdKwmr=y ܶub쎂n]xqe3vx3m#c}'?rxtq;g7(WP ɔ~.ԀA[c3(QyѣD23DYSqʌ Q_ piׅem74vѶ|iPΜ@4\}e˼=Sy=~^r7c0~6NFцP}Ƃ2!q[!7M]߹Sz$|!J~"'ecpz*1x1Y%7ޣa(Πב[L"Dɏ,|^@&sEѕSwnIL&wV7ZGJh8rЮ f495z@98dQ,+ ٻ6r$WvR|30nv0AI[YIr2-^"%˔$3&*"dYqJssII$J)q -{< `m4*2ok B"e.{"- WJ:69yREB*dV{)R2WvVgCK\>:f},k u?w6.?@|-gHe;GT :z©6zAfo%&1=_}% :VɏWeVͳ)(fɚ\3t,/v,N<%-?T|J5=9(>6+y~|ʦǧi7[:DZ^nLwE3b.CEZ;޿2ʻcTZ&Fd%+pRN)%6t9EP| wAȨ1X*YrAГ8L.S97)j\R8W4P,ԕPXxT,HTmqSq6h9.Flhq4|yg #6Jf3I_I!# +0 (V@SS"$ =b( AHQ`SjhHde.S&#VFjF0-TPvڦ2j{hI6xPRz!HiT`I#W^Ms+` Z@"3U0+m<bI@21CdEL&&Z$aQq$|YEdT'k:&n' ~wU=W~¸#D?z"t\Dm/&7w~_E0HP_lil Z7чA~aN:C<$ QڤQK6[VYb2V6L,#H吒uv+SbQɡt6m7۬vwy ݮ7ez4{A@"Xg֩蔃dK31& 6{ⁱh#QՍDiٷ;9}@SHnBLNDEm1X=j`&Ehίo.abiaveJ1L/dع/$RvHKRx"a}otj^60nn]R^^Vp{EQDӈ`e2&X&pubnoN٬h:\V 椏źa@\OlS_w\l\.fsrnC1Fi$FA2ɝN%3rRW,I#Q [ ّɝ%KZB4-RO"hAz1hWn K]M3ȣ`EḋHhrJA wRdܣ #_+w8JWzaNL((7,=[nl6hs_9UT\#B( 3brinEŬASx&UG'! X/:癶`7/Lߕk>h?qmDO2=JHՎjoW,?I|Je 8\_¦ni0\Ad:~ԮmE>aTƔh?Hf%*ȏI.D.ظOpUM; ;bGKX?Ƿ:Aknit`XȲ}7iZ-gpsїߧ >uȏ|<_ү˗z1{7?mogܝMfo2ؾ0HCoh<78i<C.Ë=yyRm gbNHw1)tgBԞAڝ+˫0]ANƑnb<ȽKhpvd& NFeN9:.0OL> Lpsθu19-u~zї\7IGi/Iԥ0~[qH'1&Lt<*i ~ 9:f*#WSWJO?烮?5N&?\'d B5w[g; q2,Ħ>`D,4}Dzwl<6ǻӵ?"EUV iR2W!ƄĐp{&d.!hŝ:W4e3O驤+_zSASA1J>$~zR6(M6.jiJq&$뛠AUH.$ygY}gz)QR)"JK ,ЬJ<: EK|]vx^Z92b1<ː 2"`::E*& gX;tJ<Hv3 =* _<[KJSwv;l:Wld:sq+#vJ(mxa|2fFRetXV&Id $`7 2SJz=*IO42d3 +HJt!KvtbکL\xGɢ1Be>BJQRJ.P(X"SVm̵} [}J= MI ]YY_ ++p'=|%rZCo9oUIƪk.{]#ǡʩM@暠kPh&ٞ>'~> 7>O :R-&[5Z[ .BL4s 21S ʩtc'{!syh[rH6! Fl#|Jڕݱ< n5=o.]~n~O>;@w;#kB3Bƻn-]X|2,\׽&@†4֯1:[Q-^T^cv"TdVzks^Ėnmyן E|#K 'FWk,2n\S(Qnœ Ywwө#Pcd6fiw"$Pe/BIR(aT W.W#@ e )q( uےUc.tWg$4=/ ($~S$bCKJabCuri:%ĹТr|W069L֑1MohbV(Q)cRLo:ЬY7[BYD )$9ci7-:P%RsC3"yIKUf=u> E:qN~Naˎt;阷y"K (jS0?< ҴDz΂){֦p̼21(,ń196Hu'!-B+|xT@ 2#I!s4Fzn*kc>h!#>[e0N[Ǜ)f&߯mݝY\OoE5Y@;nKn3ܫ ^ϜN]4%8=wfsOփeⶡ+n/jE01Any4jpmrlރd}2|{Ǜ;J #]w #qs?C*9ƓV'DSeGzyp99`ɮQ;ՙh;XJ5]ko#+q$>d Nl]b0-CdVC>dUd_FP]BV0|0ja>͇"㡚ďÆA}7Mwo};.?~~G-pB]X <?꿾6z5oj6k=mu+Y%>|8c D\J0$ep /|jzYOl4$a ]A2 dNaQ)\ 0P|Pwθa:/Li0jhSNtN2Zi,+?MKO~\w)|B&$?qc2`_Z>_F?4d%ĬZ(ѿ_]{\vRFtY 冰-\pĎ#{2͍K}ftNsi] ; ;pXV1 T1 1qu-jo1PSq!zcW-f[UڸrC4"j&A`=w$J4 2D!4AEO0c}9͡VEZWd;b!ϻ}Z=ߙ^t/ٍY_Lc+q3()ƱH)<.B/7__4=ȫoE(1NF/ MG݈ΏYNH+x\OؾКGnܙ7PeҌGF^Oczg?QLtb[C^ArIĂ> m#tZ?tD4Y^ڥ/i57u4Uy3AͮLb|:i^0xݷf X*{݌GF}y88BeӰ,O'?Zi8 [r\4L@x%Shɪ\&,r ɆqТ zƸςWJ9juD΍5]Ahi!νm=9!Y*9lJpQrLǢ*eWV-I@ iDyDq 'czq%s$sLic|qC<9'‚3L҄kRr)i2m"18X qAZ&s#w;+U^/ô58k05& O_bKtϷt8 |9OUUYU%UI'Վ@YE,1lPz/Q0 dv'T sA-omtY`hu.jg4փ!rcJL@V"߲58Y0Ջs@ *\Kbh ggsP Cf0z p!z#bIŃ47U6#ŠUf 7$e ĘU٩$6ysZUVGٹECnƧ8fJ\/< S1ĢPe+RtYEnj U[lD"V$頄 S4hAǝҀO~BV %;LXX\[-WX ^U4 ߹*. L2^H:vBTR1#ZfgA?om BF$}]A_ǭK$0^9֎4-`xG`eUgG\yWG# J}= Ψ^˄81xWp|>l"F%,B+HX^C,BT'SLWOv2=L:dL V@pIee Y SJ\&/!ko/вufF&VR\H\ȲR \gz? EĐ M9r=Y~ UՄz^w O?>;f:9>:N:9M*T Pz?nܨ^8 DޕwN .)9śjT`qxvFMhf6ULY>Pۯ_hTH=Tx|'Qk/·sSd>ydֽdW&m4,@XY ©d8! B~1Υr,8d׬򖄮 Ie2% L/-3K jK J/-ZJv0Ý+tkJA)DOW{HWe <ֶGWUw6n \mBWm=]JYOWφ̚7$w;avLWU;67Cka7tJ]te6+}E/D;DWb6k ]NW㺧=+ɌdCtEAv \%BWm>p+!"Ƞ3tU *h J=]!])\ui ٝWaW誠R._t?t:DW؀ ]Zv ZNWjr`[H8݃*ҕeY#Ɏaԛzo&UưLJZ/ΆLw% >AKg2zl"YTmcꝍUz Z6W/%h@(vhiYZ(pwB|#Xۗ J;rFCtE%v] j_A=+[4d>Fpwr#VL ]5Eoќ-f3O5n b6C}P 2Е꾢Ȥ+ w \)BWn=]h{C"KZv`3tUj*hNWR-[gOWOBW.iWX ] unjZmGGM ]^]J%{zsU +jdwq ;RfДy1ԘˎV{xaZp~i(?տ7:?T.UzVkތP'}q",+C#|UZ_^oxܹb~I.>PY 3Z'uEJ߆ ]̅%okJDl{Q+yŷ Ɛ?:}7/<4R&s4OgoF)a2XKMeuQQqADr:{ߣYvv"ms;iLB3bxB]SҞ^/ᥓIbrΪOp*]>JyBaN\N(<ނ'Aq@Myso2-k iﻆv'4S 鷊AiLgpiYu0{C0o" kcJWV0QFł~XIhԄ)Lj&PyERdI#lX.G$;+!~]R"ufIrb[+'\fAR`3֤R*GnS$%c aXjDk29rYB/EKMdHش#)TRJ&+93EV s%=iR*4≃KekiLRג]LkUV4IYͩ7UЖF r4: 6&]>{=Sg&BlTGCZZ/>rCuD*i"&K~ٔDL"a4ש F#H4t1g49lwmmKWNh"  23@%HpX, >HCٲ-b$HqsuwU{v͞QG hJ#IM^m!M%uY(/`T2XKwC4 {>7B<i>ϩTRnm5ljxIl1 1Ú4 nN';Yn]=Q1c##$63f Ȩ*BPZ |`5 ju!BRucoP D@K`1vҼXm6:BQCBNFG6Tق1K(# Xg dGh]udV@N`Q9 _ E*UFr>90`](- O`fR{|WZd\R6LJeWz#<*w%,d./V [Fhcn eXLčd XqkfC'guc݂3?K+5;xXUAWٌy4mzdF)C *TD@5Rhٻޠ\U6#ZuA))x X#A*+wN dΤfC2#l}o: ŨQA"R}W:a UlAg7 Xm#d*zSJ4/9fH57]Ґ?ȁ;16J̷5PzJB€(!2iNwQU0uk)& AŜI' sG!!.A B*)U: p)9mL6 _ b7۵XQ; )E4GRF;U B!I{֩NRl ek8ui06Օ/H!xZD5=( EdbKRt ueFjD$J(k! e 1P =(aL|<U+cG=D &23/7nz .USW̆KHN+ .*K;L'Q_ }0` zqsWK^\1xW TUz@vM&!jhX 3 b{PT8x0(AOБW%sɕIWo#f BX9Y4<#yp,_nR( {򠷒!HDNkT2XÙ.hZ%f0A 9ɐQgk#ԭHw(~&z㥇 *YȩՏoTG}^ yG;ێjP$ a>_..C2"d*S8r VL%V  XF;KrIt ȋI TɣB- āpaFH HnLCQX1z,Aⶦ3*Z]lJvl 䁈:%xjhjXǢ-&T3RF Xx;:5^&^l#p,ALJ3-;kʐ 6"QxTDΓiz3dNғ֞fJOQ!,yJi]0΀n - P0@14fs˥ZpUn\CC,Z*cQ4k&Cl>L](Ijl$dipM(sAyZP^)T1vorj i;GeOf,AXv=Bh[~Gq>O,I=2!C,ud;& _yo>.r+)ZoAkjbۤwO _׿܍/~ћ/sw75z~{}R>CK+KJگ?[F*ٺy$?>>^/M[/׋?zh`~}ÙgN gq8sq1DJ8N D(o8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@Ls:=8XG8SGN  tN'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qpS( tN I@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $Nup3r1=ȱ;nsq1 z/f tBN S'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N q/[t=iKjʭlo7 ޮuus~ b#2|KlK@2j1.qOCj?mq:`s+FeN m'K=<.ynCygFn[0PtFLſ)X\_Ea٠g4>Eޗ&S}ۍ4#`lhZ=f;M3OSŒ 3+gCW69@HWAi͎=(4upC ]UQZ ]'R@F=7]=pTKW9tHy#QtTtEjBW3TfFtI͇ ]1Zwt7:A2d` ]~< ]1ZWBW'HWaFtA͇ ]1ZozҔsZ` צxt(S+öT2up ]1t (mdM/2埋%'~;1zu_޺{i eܨS6/Wzj׽~Fp֒OjL sW)R\_O$c>q\wwMyWhڲeKL|Tן[`T_5W6kh/+'5_߾ON;Fߔ'_/htϪ^vUJZOp4N%姌Lpzө[-[2jKҡSiIIpN1i-vZ ]K$fj-k&SO|̓ٓ0'wnd! ֖/SncWѪq.vutF8l?ZI@uS }]nPq.thˀ8$tutetq=#`O{$@s+ ;]1JㅮN\"ǒ' ]nг+F±;EFtRr4'u>Ma.tc+F x]&.>#bGW"I?dveUJ3iϾQp|n1Gf!M"M1Zg߻CdoOM=0ӸMmVrD6Kbsθr? ygY0^Zrk۲ncs`͕ﯴ7/rl}s*jwov;  >:?ܟ4|n=wYM|𫻨޶aR|ĕ65"G tב`O>*{Y;vs?v]ΗW: heuN&tD|m(5tVyTm=KR*Lwr+``ÿ5 f e?e/>}9tWwvFܾte;.j83ߴ% ؞௫7 M^] |i4%rʲ_ͯ_y蚍ܽ_|lR}!N t{fn#gj{ȑ_q:K,8pw b1ew HQ$$,_%ٲcYMrd"[SSEcPD#'iCԤN0+%IElJTu|Lw{Ȃt)$p9) /` 9&-u_Z~;G;&c.w4zy!&A/u`09 )]J>I_,Df'=eO߮rG074qN0 ._tFFV:9_ٺt旼} 1_a׿ܧ J-'StJ3tF֖T&(la!UTw S?LG~?SܫTyt-ǃ4Ph􂲧 lֲ>it5^.6_Wb6oMCᚖܖ}_҂JfCЈRs^,dE@+'+jW M.e"ƝTrEn}Z(ti| rz@yTGjy|^S)ч&8߽;zgtǪ}F{a7 +:W=f{v7؜ydi۲^N9mQхqgS*]ݞi:p>\?EhiDڠhAI =k:h((Q*Vu+J#x9 nM8j(B.v <(wefO  ]#9)"xYp#b)鲀$b!d_(UO.J}I$1{[dTڤc&Eh~. ugf>M/6jz,m~x-:.;E綄/-{۲d6Ym3շ(S)=;T1U`p9*[E'=٧iu9g gbse(P,4w tMJN#$?[Oc/w?/n%DR9BVR*dj0xGhsNGY%ARC[3\:1h$PL\Ęrv^xWyPgIXxC(7ugݣ<2`k/~ 'akDq\[ǯۃսyVa?? !*P *mR6D,IK+"`GV-<.KfN+Î5AD4iU@7~'f]ngޱ)P㭓̫M?H>17eҽx׮]Ϩ]ێ1K?vX2LTFaڗv:h'gGlA?D/KIP!ǻߕFwUT۩_,n> &1݂<.w;](,ʼnjŶ1F 6HWrBAZWQiބPbN4=L-~$y|TfEfo\#CF!U% 'O~@8O1gBJb L3.Z0&;-<'Sh"г֥}eV\dt Smу,3HJM_,獻s8a1)flo еmވJ4gGiX%ԢIo8SUItAtTJh@/c<.Qڏ9 D),0a@ΨC-Y-;YXo sEڜ!砄 c+m5$-v kxjd}IRN6#hRtN ^[lL{GߋX)*OxDY edMI4o:>!PJDGUSdJJ%F& hz2՚5.U 9a)\NVZ4 Tkdl֝؎4fT,XXxݣzmʌwdw]/,h}_Ѕdv5~WFl.o *r D4ҁLJZ&&v'|j:FUMmYS>9&MZ/ |o A2^h՝gӴb j7kn@>AD %e bP#8K;I,4EMt@Lfa1̒#df8&cMʂ 2qYI8.$Ywao+0 "6""6FDq@0tQVR-^@^F<.Kxv.clCbɁ*gbgf$#@*- 3?ȸuT[go~xAtIKPKIC$$ZfT Q3Y E`h\l5. 6'mpޣ *_#o߱q+ܠ~|#dnfQr+}?ko{KCXW^ ^oHMф;eՕj+Sr CQ\fGSj{r4W8Sz[%`ARhP*NX_tA%mT*jZb>ח_J ;t.\.WT0sM?N8 ľ(Kq]tIxV$!Dݎ@&,$kd"a1rn}zdfo9gܐCRa'`a|ug˳?f˲Cpvsא1jBlw\1LDΪFng- 'cTD݁.E%4f-8szwc d;K$ &%d3`bC4֑ΙPG$ %~P:&4L-lyGЏ8~ag¼^n`PbAx Tf,Q -dS Ii"84݀/Ֆn'0 8AgbS9AȔ@)EkYE+! @ aҨ5°`1cBwDّZG%ZKNȑ>zP!hR0IJ&c"RR@po7pb?ZJuQud7;'N6K-@:bF71DAl-ER)4\bLEꜣ"O9єh^rnR(f1} L?T)p9I&4Ca^?NG~.U3xg|0:_7/Y&FW_Y*,6YXWx5!rGeZ,g|*-uo!b~XUe LuV+]Zp|IӤUGy%ZIi}]J Zs 0 ivu1\5^=gY8+SImPu7M|S]२~))6I3#!#@2fZݧ_8?aL{@~{k’Ϳ|!{|# _s¾\\L\ʣD]?Lʴa}瀭Btرh/Kt!0zp@{<^^ 蟜0bP7^VU'up4<>{nK@.t^5?\=J8B:JQB0EO齃0&QWϧTEåN]dǓIgٟI=B5uF[ @5JOp!{ɨ:,t[V?^67 'ڞ-U_*,MQ:ƝY3RBS|97|$ozByM~㷭>P#S8 _: ]R+zSgOͿ2,iIL*9wI=Ǽ2Hjfr:rrM$8$e.pm^_1qqoUUvTKS JCn44# CldITj9pmS (WIlBV n}}T7Ხ_gkQ lp|\lJtJb#+9=2NӐU.RJž'TG#hJH4""#) (o93(F$KUA"u{)b,5?==wxK9Kc#vRRGG1WtAG8_]G=a_Vϕc,\L_r;~j04G9Jh&4V)_Þ[ 93#{Z M|Qm84@_ LeMq6uV7OjX?~߼^|?65`B3:|rZ9VbfswV 89<5lqCqjNH>E0}MZj!jD!t˫1i:*AGNmֹjڳQˤ Ls8>\:Qb4xgZכ*SgcY_g?<??~sǿocyx+0^+ (4}7OZv547*v .g,*.2Cĕw3\] J _duڗ*mzRWKI? WF\Pq5^:=HJ :eMbmU#2Wg̅٧fKGp$۝w6!/'Rb.J.?M~n@sB \ok}R/mw)r7 W` 4dR<ʯ *KC_崏 $` "+&ZE}R3Zk1oyu{Dņ{-jd^/eB8mx5;$`+kwơNCٙxҥҬ{ǘ6gfn6s [d@)_Oٸ^t $ TXs&(-:҅HbpA'vx \vŽ`e` Zk Xst< `d^*$^ם' ̭<(1)bbI,uvNv(ȧeg^ jm[jr Ja L$j*(ܥt*"L{gzիu{kT]~ }|`*nMEZ{ۚ\y,M=^݌eVat6^~T} Ywk|>95} dT59iRGd^m4Z{Xث|y÷$Ґ[濏v`M3*gRbraW*4b/YA ([£g}Y^$ (@2"ˉ\"L! e)|b., P/Aݺ[$1G7C:4M1CWڳn]5x-%9 OpŘD]jo &z8S;/ &U6pwkfS)["7y$p/*bnb+}vy&u &(a*Ts-c1&N:tx R!f1O^Vpb#Vt#VRb<\ {;c6s|M5ǡ}j=XGziS%O;շa\ IUpR1qV{'+j$Qob%*o3\ۮpUpb( Y9Kec>"CD2S"*^-~}6cqV >-&mX'~=#*{M^L[>=}cwWꢀԿfz/&̘7;s56ۭ@w3zR$J7<5126<%x5<'Qsugrg753us"\)73EAQkNF MpN'CAD!L`cݞCQ{0r&V8m0RS-y"J!,E N<<:n4.>/=N)4*fViQxP."-[eJw~dA6馐~SҜ\b -ZgbZPB9dF `*գ3J¡^S :q)M%VL8*#dG(o"Xy%߫M CyN;N LTF(`C) )5'D)&#p/A۠;+tmwۚ%Ƹ]S>:PaUf}sR]9k3I; פ M IQdh^cw1~UŃ4֛|;6}CXN žJ#68R$#Z{1[,} Dݷo˱ ϹGg# Nv7z&NK+C4Rq:aenNb-;/i,ܤ,9Qln%9 xޱXaSEO9E$|28M3G ΢"*iEa%VE&Ey)D\h@FHiC2 K D3#3Œ&!D!e!x0kIDk1hFs+%] M5n ?˪M_$;ݫ,rk- Wpr.읖Ҫ6d*z׃\_@.PqISg0΢CZWOa9Ngm(ݷWZ2oN.[=_5& =ϵ\~Fo\y3?(jW ѱ7.=PPY/MWh |S j_ȌGn쐤ym'?lB(#,ٚL \E%c7gbLPR31B)&t3F]%ptj@ɱޡ(-BW `]%pם9;tbL6]%(.KDW"$1c[bk'httP_"a,]%tƷ]%h%tt;%+AdT(G՞2`-*AK7JPzJ ®Ԗsy]Qhor^ዓ^v8g 啥eUވd?~㺩Ip4 Xx]Qtuev]mPW#HWNApѷFh5u86hOt`Jy]>2,]69YrmK^DW$֮8nr h}jJKuQ]%!'q]׻aڍV+jbu#%ZK<6}}R]׫{>߅pT%O_n!>b09FӆK|"j_Иki&5syDpSEWFzRw]mQW38 8=pu@+of}=3n]-kr_{72ڨ2ʵG tE^Iҕ;02k7[Dz{e6;HWBpUGhcXhɮ\PIW#? (2ڜ׮+P84 -i]V}\2-1N + +U7@?xx]Wҕ*r{,ođ!Ҵ i`ޠѫhY s#ֲ%kkeqv L墣r>E͙~А׃-je>zpwoqnZr+\",.46`g0@7e{Sl!-ɏ+Ía]mZQjtltb3_ eʟ3/)]- +UZMC֐ҕ" +M2KF|];uA] ,4>+%v].hu"I@K\/z)] ]_2]WOt 8veGѕѪ]WyjJ9t`a? 73m eȻ6+HAStp}A2Z(w]mQWm=oYBvɉ|r eZ1LSgl4LVik״Qj5AMTGUNޥq&;ΚѮQڤ4?82\h]WFGW|fK,+ .kvp㕏_FW2J]^+uԦEy( +Mi](6u" teѕ0-9Y2]WUdMy ]nQtZq u""Ad ]u+f&"!~etyuDv`IpEWF+v]]WgCM9{w pi]mZڕQuE]ƣje';\ }R[ބ< }]$| x՗,*L$E]1Jܻ(W%,XbfJyjYȬ]-Eku2073\q,-m^JQ*K \Zș9 `㬄n v]+su@py]|F(c]W_̦[u pt/pjmhu2ʵ) t%^Z02h)]WFiuTʀ~A2\FWFy2J]W+חӕ8Kt!]-^WFquS)2`gpUFѕueQw]mPW]Бѕ0ٕ^k"Jrnuz|W{|uPrمt_nQ* iý-h?ͺit #}" * h]n_WFҮ-Jxnsww1&!Ifmw'<ʝE+9`Y 8ssN>COJEћ^ާ](bO#v-Wk~?gY0DjdM(XS;_߽{~o^=} ?z7xj澖vˏ79@+",a|rLϿ}n U%Ps+19ǩ$?*qt7N3 ?Y7?}x? )gS}0GYn^?K\|sN̓Rsp%0IS&_\䑋:SҜ85ד~h=IxT/-FW&2:60Gp4;ϧ*ۧQ/WB ͑4Mhs$: XPߜJ(^Oԧ2B.$^zGbzkVujh DcR}'_Z4M>IQ5@n޾{TbiSjS e>, MaN>ŊL*u!MSͪp/S?#>0qBgF3Θ{D3r95VM-TёFs8Iؾ DG0c}:Qd? :j :yAHD]MaBtPuFS0="A8:*2s,QfN͡}yeD!D,@L;h$Szߖ߿)m]\-Etn֊(֒8JڸG_rODIz7O"xU͋u&Siϒ\PҊnφffHnV_KpeI@\5_҇M}QEKTrJ.1 uz~,hrV1LA4]qm$DI_"DUc!"}BbN߅5%ag֜ڜbGD'Q0 t$ࣜ=fPT;_+2̀ZE.OqB7g̱qP01{v}:!Bڥ1y\D!{ü+V' LmP &DLdѩ D3yifԂkbϹ͡bZk)IXYx/PM;.kGb jӨ6a&@v(R0K@N&S #،) 6Y7Ma9d tٮ9/Ph͒[0eC?c´WST1!avC'ՕbDTͺ4H{i[sjUV B~RaCo'SĄ*Kec`E T$l.pАxnb~{\4t"yffI0h~yDXq&b6bplXF5KНDNP?M5dwgy`oܴ7?|otGcV0Gz=z8ü &%{6sVX = xgY8tN()V%KJUҵӜKdJE,-0Ɍ1&i$|Ea1 fO 0HԆI&ƴ0^Fv/wtzmtYzO%JT_ )k%ƌ66x ) qo,furLTG֏Q?yˡ+b.NAmlrj0/$3!OEb~~ѫ?wd$ɵN(ko^allVH|?+Np2ާ+ȓ.4N:t+`raBo((BTUV(LF|T500fg p l2ah*~,;YAqA)#܊͑ԐcAT3g LA &}̐H+0q1?]*Xew;T N>śNZAs8$*d!  A@aNXJ0в R ژE   i1%7 o!K|V RO連 T)C'I%n]Kx@ \4 jk6RjN:a-XE@I(;MT!2l.H䮝4\F @/ \c#:ct·+,;70&`A#|.~:ZܪiRnm- G,u PY iYVKCrz5,|xy /qm>Q<w!EhbC 2(拴cpE:i?/rx:?\tZ,6)t2MGG xye.+lfGMh^|>߯2(i_K7ۃOGli6EcK 7«d9~{U8)UQ!\sB‘!*qڑ@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H p@>g~G'."\zz0C0{J դ"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R=\%.}Ri1]@W(4{%fqR=D%H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@zJ u}Rk%7J rj@=EJ̜@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ^O xURS|vy}^_\?_P-+=d ]s`sFiմxk`u ŕ98:/U<<^[{=P^HrKsjyFϋ/j?ֲ#Sm˻O[~kp04f|~:#4Cm#飿q6L;Euw<֪t|sXg p8. ]Pv-ڍ% Yuڽo8[9@k[@կh!`MY=?ggi.T^X\=뿨T=:Y)J)}Hux:E-{'4rv&] zCWWU{OWR[ ]v=?dzVlstZIKJgtʼnnxh]ةbGthDWMo{DWHBJ6NWrĉDW_]~vEW7NWrf])s=+EA+u_ t( ]=DҊ9{DWغЕnw ]Z >D2J!{DWX5y3hҕe\Tg'(\,~[#]dg5X55i#W}[%]Xv#0'b_'ži0-xUUtPJrt%]^¸2 mɫ#u~bn%_ϛ*kFo.%Kɴz糼 kaH|=ý{:hv ˪> *3~;zOr֝dgmAy}i/aq&&R-mdgH2k ZJU:S.h_^,ATQj'<́֒TWr>*8<^lR[.Lu,`֞Q`t󟎽h%lS0WKɷW~_^wl<bklTޝmֲn`ůaz,G$XKhS^h2 ׆VUt&}x߆QhC\(1,|5/1WI *k8x0}m=g:A`U$Ik1*_u%sɋ.E}-Mš uA/j畎Y-` d#5 pێ֠p|uY ּ s/= α+6"Lݳnͣ?{[?ŧlIkhRxg2hrotT(w^̀N >.SWO \Y<xFc{lz2x_#\%n8G7'.ҿ*ϲxf b$;GM5u?ZtS5bs5볏VͪgyZx늈Di.@R*~{YdLBQ:9 ylE/vrh;ocl„tKyMc+ndAл{#&}ڛeQnfwumݘCrEۋ^8iq.80#絜PّdŦU3|fnˏlӂ?f2fV)tveBH.2p7ro.nz1Ike۳yRj| oosi_:?k5o ^oMQޞ}s*]kXZ \2wrrBI>ܟ[Eߚ"H?yǹLu؜NTd6]z SFW):(s|>vp1$)PrxR2WS9s̓p4#\|R_^pQ)Mh$MZ +[ACv F0Z|i?l:C$W58UA;V M*VTmU6'ULj 9y-YyY=TrWqs )!@i[ SUj]R\*ҭ;j jƢ3m6U"TTT\ %E0USLF&el(FHaҪ=Im{քb;YW)[ Bƒf]qhRLhcdoA(F nb)~݃rLGG@P7;|8فIy+9%NF/;ht1 JyI|=9+tǟ xAAb]BUL{Bh!l'SXa󤊬T_,\ *pIQH; tʦ3֝1̉Wc}ne3֗CkjJMηH}˜5gXhZ3W7|kx>ß3Lcu~kpoDw>CdpbhF(&9T+`,8#+@q v1%zlW8j|_?4 lD N80~2&:s?ꢧw`c~[6l-۳:jr:_Ny{."grI.x[Rl)(,6]`j BT;#[V1چM~;V]˲Z̙|<;ڒUք 蛳هj#n Be~Cn3$O_A_gvnN仾~ӠnýυnwCz~;TAx۶X'w$gT0{ò-wҚ*1@Z U7"ۜ;F5cӚvsg/Bkj]BȢh ¦ g]mm±`TRg!a^Bg^V5Ze%M9S6DsѪ fDTNg3Eξnyܷ8tuy* zqrTPR)1htfǘ!$:F֚P0ete Ad)[J謬)W%bʔDoe!J59LdR%mgg؀PfoUr)k)4 dMh[8X-E\ 5 T]!QtRKd/lWcuH'!&'G]L:Yk[y3B±E M66"NrgO]Ѻm% ܁Gr\=r_6w|3|+ZbU,c4@~]&Idj@ovSV٣AN-d FY餫%`UWXlɐ+P$Ӿo}-7\^^nvm'y֑V5Q'u8NX.Ezks^:5"{Eu@b_|i |H23#,tu0VAUM>+810BR݋:_A8G9.Lb̬-)(jujLJu:.e+_PTu@}p٨IGD1y[w?4LzI;/|HԲƆ/rX9۟]?+\\j\͋ԎPƽϗr[;PàR\!j0C- S1s!}"0y1F'QBY+3P85ZVɇ@ܴFa2JZ0$s()())sbTH#s!ЖJND2kSU)1TmkѩR6EG>o[wTZͿrEbYZ(H>}-WS~ y367>CŽE .G^7}f&=c{wrgW/d'çԀ[pF!?6كvp{v4zwn~yAA9s3hݿ]xK?-{<\_]hout53fu?gGi$x5{pkWCu74W:u_ؖL!7yKU2ӗRvX:sqyNŚK3ҫJhɷ# $]Tec8g5S`f2h,*)Bmbg :W7YH";ДK65)UuR>*PCМO΃䑭HdV;خWwmS$ EokQ@QãծpVC],l0 dcKBlfMuJ[lgݺ^U%֩sʗde_<@)iM-9hSV` z1񿜨B.ՂQ*ɖؖs!*U%ˠPDhKjo.'412X^|HeetEfbLr:q>ǜo>KYncbleŞG, ޡkK sVP &Ǣ{y,oD5O#"}|C+ !z>6Dj|V۰{~P"")$/ZH^šն2^̣y܁;yuYǫr~5219eXrPZ&8Y &ɑ&\GH4>&o0\)3 p `Ubp0dDGM"G۱ijO71R 2!M̈́`M~*X^X-K zp!bƓI4e $9j4";D0yP1d UPӁlȃlaD0 [n9ˑ!ȲY!En5wY¬bX*cM&Qྡྷ]Y#G_5JT +Ggokga<2\c }Zukw;JR5nt%JvLaf=n}scĀWW O(kt_ fFJ[ƞ”fޢퟺl-\\/1|L3Ė oݚ ( +CHPBz݇&w~;~R_쳰KlmIT!*b虋 Y7VзWpn-2X` \F*C`M -k, MؙwwΞɭI&RYl}^Z->\ 8_-.'[S}=X6~TR[/Zގ\U-2bP 9S&P2V{E p육NVytiMԂ>5N !H; tʽswF| Vܬc S >jk'=W}ْiehtOt_Gnm, ~er{s˫gZ:0Uoر%7`b.遲;*ebbx(KUCtdVw(|xŃ*.)6Ț󇵣Leݓ?oPpq8xO@< c!Q#--4A{K *((UAJp#{YM4(٠QiĄª1lھhǣbb jnuڝ1S4Zx$&11{THaNf-ə`2*&Z@ 5C*Њ5bM̀B1WGCd FuԮʜk~* bk/"t!FgpL'x&n BIPV X*Ck …@8JaRKrUTD"ȄNS#m?E\jXpTⶸd_\-pMo#ɱXS M$ʨQJPVЎr&2ΈDk91t . {!#a߽ lYbMѪj=ri೸Ʈ+ɞvp? 't&י|#SqJ Ksxlso IZ&PV%cE23p6L¨H&&P‘cZ ]+ꎹd:!kl }^Dy5C׃!bvȴCهG2::N2"HSj2XiE4ȥ$J4dPzB ٮ$X۾Ϊh)U:"hoY #X kI!0&v&i5rKz_@,zl嫋3,?e)埿}\{~RykߏZN?OQ)af=̺{/ҏpʜQ)(PE%󆗂'SzCxSTR ^򐏽]F a>BMj!ڙڜ|Xi "0r^* j-] "pF5e@w(GUfoZ!?NfvWޟuٴOwl`J@]kQJ,QG$8O""< "d^Yí>!pE;q20#QՂĹpyEF tu[>B4yIP*ӌOswwVcMS~ Elxnxd2H*5QlAD3*$ɖ ! &$ؐƳxtv:c-tWC%E1%eN+aAƹJH4~ML@5Aw du0yAwgFg/6zQ{@q&ty9楀d6HJE褾5:X~^W-GCVu/mn(G;-Kmx.AYCa;Oi8*jϝ`#%B J10=(2'^jA܏=;%g*YXƓ >wzof k%Wn5 1n?MFHo56(4.4[45}Gk;~]Yv~VJ ]+ 0 -Żk7OqG/NzیϽ.n/X<;w_'7`#Q轙o,G~*(KƣfWTyHd:a<#ޯ3:diYkRrĨ>G؞$D?Ũ}2ϟ.r9#op cKZGP. ''x[Ζ1(=OȈ*rK:0$ WrH/qiLGT^:VMw_H x{On]Hjǯ@M]ux<;Rϔmkjl8ܙgO7eXe n G&1MW]x7PԘ{fjK<ޛ.`cd&i$5 ~ʜ\ZLotҕT}oe"8-r.eSZ*XQ6YW#CKr|cK{WƬs.(VL\T R)jx$Ni#3bڤJJgV8MFU ['AS`M쩗W,o.Go0ǝI do9@,^ py~{0_>qIuK?F+~őA>^&TSƒ]%0Q.1Z&K(E.֧zDL'!'bړGMߋYZ3OIB=UHh,@&|<- 6[Rp"DnQJ< 25u1RƂl-\F"j2@ji K FΎQhz0ȣb5S2 W,7WY?D>\W;5ޥxOVd,˵n:.w 4zǐVzAJ6)6uN@API)ƭ_ul!ͷn&|qyfܛC Yzey(Tr/xQrPW81g",~D|ϯUMwS% 2m|f_hh2a#_ET'ʿ1avAk3֛'D_?zW]_Yrjp Tk*R+}XW{2Eɍ>k9jDW.[mŕy#k}JۮQo)к>20/`\Hm0u,Fh5>BtSZ,Łu".fA.c;\rcJUBڱjr9Z㉽9PvX=$7k$ntC$>K' 8?Uz0=^%O Pr;Y I鉷Kyx y8@{ExB}bx-wnm"^9JL `4+׿EvyGhKU*,lr˕B [2KʾiC_h G,3 >K@-ITju{2ngiu<r5֠+&}**`,؁A!CJ*{j#P-*=aպZl P ݆2&YTrͦ*Uz[KjD h)y'g9y2iA\Tb8*e!@#$m,iYC`B#;_ׄQRQ H83,Jg܉sbaևrߩCx˚Td& ^+/nO޿,?.a|khD<4NEI|(8+v_ xAAb]BUL{Bh5ol'SXm󤊬T+K's۸]=7FRvrݭywֈ[nr!28qdZ[lTL`gde4S>λFo __͒PbξNf. |;ǣ%o2zҰRX@|;WFkCKcQq2 y 1;R ,q1-Xqqlg4OSnĿ%xGzٺ8x[Rl)(,>]`j BT;#[NE7ڰ-y+'@v׆qFu|ٍז2䳏َjs z76G>+s ⪇9^s|5c}o&}/2'v;!'nozX|rM@̘h yh Ct F6'Nk #5c5>XEuj@ 2Z">**u9 ǂQJ͞y c{YXhڐ1B!DVU0#r:$)2$.FwuӱM ߪMH{)4ѷVkeR*0^sU1U@@^Z L9 A){J謬)W%bʔDoe!J59LdR%mhoAˡ?}ftXSnudMh[8X%E\ 5 T]!Qtj_suH'!&'G]L:Yk[ QGg "*c|lD>6tp;(gnA[ۃAS8D@ /\3yq׺JK7MH%!V52F!uh5l´K>4O-ߎ}I7ȩQ"+t)[2䊾DG:6I(gw}m[h$u։HkDeV7N [i'FۉQډv6`_\ߴq1Ky>}Kp #T4P"3*gE'FF^ѽv{B`2(h 'vŀIUeP=EWEcUB[紱֥l Q.5鈈 ?&*:wqdzИ;/i 'ʧ[NԾXtnzuQpֶ/S;S&)ۻnҩ{9H}&0y$_+wϏ/g7~ݷr1yѲJ>=a&S5{{JA;0+wՠBB Tr"XJj[NR):LoEn90deA??4Y7xys4~tU+u[],_]p__ʌ BVY,7_*PCМO΃䑭HdVԻpk9tLOK퓬/UF ޫ[zAӈArڢS +*)S)lbJ٠Y b AFRsy fޒ9+Y]u*qXZ k l9ϒcJ!Fh;(A*4!Td`=U)~m8ZocX1G|I&P5ȞVؔЕG˘ҎSM. =Q0OZu"\]*T-ѱ-UUs!*U%Q(niko.'47հi@Y^|HeetEfLr:q>ǂo^a7[#@!˴0êwh-$tJA T(^,"EDdl_ .ǫ%84!@lzBZ{YMr :@BWʅbddM7586 ;v$ gk@0ےs2ZU0 #E\le€%< "3 &j­.D F$0вƢ@Ȕl4'HWЮY[ן/e>_W-wpw(|=Z~歨{/t e(4He Ġ2rfkMQebcg 6ZCbǺ&M&jA'dwɈp:2݆ߢu}37︇)~|')ٷ? Xi}vPԇeIQ-w=wF,ͳ,kLvlA"xKzl)Jꑘ*+ץءCW:{D2ޠ/p_Nt4vݕQ*~q+\<2Gbo~,tiv\"YIʉhȸY% lŁ;PqS䍛-d'ND&oM ft{+m:h{66M'2G\R$yy~qIVT!)C=[Y~6yMm16V˻eȫT פ1igX%PV'8ب@P+!M£*< v/1 5uҠC+m]0Ʈq," ~,PjPZΕPXZTnmpCYw848yG ;2N uO1ňHh* l CZ7=lH@i kʹy{Y 9_"+럄19NHFɻjД'Etm1ɂ 2)M59|SvuQY=MY6tU^qv6'3z{$k_jݻ'u' 9cwtTfBS49u@cn6A!=9Q Бȵ`tZ nDZ n%js=b[|r6FbzqfY`Ő`$!W:fU* We&jwuK<ec/b7. nP^:C1AALwǪ3Yw]Nezi*+d1`_C krwCR9*B_PDFdD- HZ3t^U֠Wޕ$Ra< mL7/LB4e5DuȒLnŮ2"2(̘")gRs]LrAv{f\n,JHL451i :gQMn 8{uFΖ!ҫH'#9q0R> Fk"R8x4 #F:Md2i:@󧀏GPHFKApJ>O_ۓ('OTp E P4A9 J`8MpA{ԡ| س,&KPiDĄcB#Y}9# Êy(wڪeVjw†hZ ֖1;T+QM3BB+GeUqN\S DABSQmAi<(rwE%j'Q{"ȄVQᙻYCi%N=h4Mk"8|us;V`WtW,Ak!pFmto 2\U2y)xҥӄ.EAHTg{#7)o'G٨#Ry ³'iHQ l"pFeBn~P1}X+tqf aݚ4O'<. 9G-HS?*Sc)yJy%h:Q!n0tK&zdu-2t0#ՂĹ9IF k$Xato !yI?T-Z=pwVcޱyjBO<pG[XK ̓j=׌s4~ML@Au duAC_9^Mt)L`|NQF0kჅ@)* m"b')֤ʱ aZm6 Gԃ-ElI_jij77/ȿoƝu/sNq}0!%\\3KHꎸa|Z:I'9UѠ/YAChlv:"?{;&9f螁hL>b'Im dǂ|&*iTMG1^zo}JTu%Wnbv?m9*z!W56ny~_Usۚ(<0OU`UUɩ6 mg%P{e|Q}g;SNG}Fx n_?CA\UoyZ uxY_h<bcy'X2*DBs%~F3TPǣaI GXLE?7z׃EǓtGԏ agR?o>Qu Tۼ_*l!ws~!XOΙY΢P!٬/= r֔1_%7(^j6'ĥo?қ7?Bp^g67,[Ҭd)EM$)OO0Q[5rI/>'yɳ&l(]5⵻$zV{'^K^g =^[jnP56,L̿b~4\.xbWyDt6VsX=^xNQf]z랟ȍh6:> .OF/;/TJ?"BeGW\f2:\e*%;+j\pX nu4\W\&kMr>C & ==‹$xgŤ )MTc()-Q(?5J*կw7\iPO^TvZvΆd5`|,Fxl]wqLm3} Z3{ CߞAѿ+qsdO0dꓩP|,B/dj}pȞ?2Z GWF+&h*SkR)dW\-C_=,arua>^F \jץhY1zDpE P"\XI9=tTZ_WOWL˥󲳺eSPlkOrWYNQ<=kAR3,D%J䞽4T2dL[l);K3EWM@/>W7(O^#tk;ϩZB? BN!sXMHkW^٩1̠z.M߼ l̯Y(tę(KGwzW9 {^Hߣ;Y;5Ԛ~o;%IB]bT;P[J HdД*$ JxhȖlvs 6FPwx΍q8D"(.(CS@X-'IFΖMhhz09ȣb(S%n.'Z6Po|9@xr譊[Bʝ Hf6p*mpoI`A'dM.NN$nR$;Q0Z ^LeQic ׄHt4q:UpI˖5r* 92CbE4, Đ$y*X#XXF̠d%Z1;ay aDM\bo^hFT(!RZP%zsR2;i|ȬA= SZQrM𲮓6 e ;@t`mXˡXk Bܰ3]iM!KC<1gqDJH&h2s}JJCPS`+m#ܹ<_Z8M ԕH]hy8) \zs}_'ry`-][>/֭NWv:A{%=xeV+EP8`ހL)WheX"&4MxPו)lL!b${I>+3bE%8_\^_O]3b-hL(ppUKrN%2Tu$< "a[ڎ6 5k*stek0xs[c6距;c8r4[.u56t)::lQ;1vnZ*U?(fz/kءWޮ/\8(^b[;m#GE/C|)۹;a3n."kǖZh{ %+2ǠRF+ ss&1"2#Q1餼{@Xs[_474}6r6+B3s+%(cR:XզJ{x/}<.gyUx^DdF|ƪdSL2,@9:@dSOCINJAs,d VtΠ0EAz52> u+Hwr47*D?(!JD@DPxqd*q!I ш`u\s|M[َ_6hd٦oaTӪH|s88Q5X/dq6AG]ۯ&W[L|S-YנS^핊^PpoFVwmncy· b !\ĜI9;],{7X d:rn~: =˓ ľ, uwnu^RL[7Xb_9>/nw{{U^׃ޱzMǚ߮Xvl}:9پM˨0r4]n_6F-?'#_2;񧿿~~}~RW^ܽIǽ|3v]uF!ӯr{)>R5ƴQifo'|ejw&~!qUuQZ?do XѪTZX,=wZ}P=kJ/hu.\6yLE]fpr[m4ըF?zyP1h T@:0YJ XRͅ།"^7cw-i>\lNaia<ҩImzd>Ig̟'U9gpѦ.^^ƚklа"ɗH5E^5 DZ6<^ t2B |`.USykNvkW d~n!JsMӰXfM&k5v<_Oo\!rrܒ1^S`zPJLO-p] _0ay~ӸUۊ].uX/upg7-Or8q4ۈ?.f[=_qlz o_t^d{v7|"u^Ւud[¿M䭤ͧ,DsYs{ӵ_m2_*bop ݎ,ᄑy~F7'f ͊[oIW7n|f-y5cxH^* .֨omݰz[ݓ9sA{s)r3tsݻZ>̰4;m 4R6TRjB*J#B*_a!Z0"U'{(!6Q\ 74x5m(<[>:N!ݞdԐJFFKp.c$&_ J ٌ&Xȸ+K(>7]|H68%%)JQI:@F%&)'%;þྋ٘ZAkIFai^)wRY,d2R޹`0XSibaO!H$E}"'r:y IWL$"FSb &,Q$ fE$QET_@P) ћ;U_}ڈLQ2`XX AZA(UUg/0Az$c̶3 'thk{3h8x -OXc,X hOjx@CWMeĥH|;y>D]tla2~wWn:`No͌ݯD0Қ-D} W ڹY+U: ,U $`7A6teS,T|X\S;(p]F7u{y(sz0S;7jxiOf-#vw)n.8yQ ҕPVUT$[x7!::M6EOf䙳s9፩|Ay"TsA ujùCȩT6)T(Նy ;)#)5ؽnB% Il*Dcn1SN(20B:t Yt@f^dlnj r*EW %PG"0khb+m'_ƅ)c%lE#GQrwCU)~VuB*I5i6BF;!%IElJe]*I#jA/RLA'oNEbZ3`tMID"גԢ`' v#OƋWw]94TF2Q6Bdd  `2h,󼸼粱sˣuײK#ˣm@>?hH_\y%!a|f °b z|)$6јD-s^:&o1r=u"w>Xe\YMʥM*hպ"RAKF^rLZꒁ_Z8Gwc:t}8aH:MZf:_foi[6<-(kiM{\g=E(fƘw.<[5wry/YA+EV2"z)8 hE K=<`fԛ0g:v~9;v )?O'ٲo)5. ޭw-!k7TVH>K܀&pܰݶ^8BGZ;JyZg M)J.s)Y/YDJy~<4zK2d.[LxMo~,v`!y ɚDDU "+qI]Y` 0vc=6CQȳJkYtXFPG쒪$Q%E%1I>2#_Fd!Bk.tۉrʆ HE Ip) Fk4^9)R2ܡ 'W0JDY΂IgLֻOs.ED,"+!d`$PO-X!35eZisY}js1 ]b&YG,죵Ð8>(f7ԑ V% {[,3YĜ -N_g_BqtrOK'cQ)YQP3= 1hUKw0h:ij6.Nbsk֩ҔXvTsE0cڈ)Ã91Q+P/uXf!)!jS`rR! 8F39RYԬu灉rvRvJ]?Yel{c|cv!,<+N`t蜨1ggJ`1BpOԀX^ 2bhf 4$ah%7,5e51κlx*Ӱù/n0}P#3v IH!SkZZd`fXړ@ZȨCS0^{.2½g!mr^&NL$!e0M[uh7L̦vmq(kY[wi; T#2"`80 ߑJ@M`at0"2S+f $2hTx51Xb&(Pq:c0vu~>U%$Ʀ0bmq(#t1Ng"N<n L .8Rxdcr nKX\`* `f)Z<㣇qTR$ A&4wE/\ur9[ʋf^/vKrD<;A 2je QΤ3"j‡Iǁ|H||x [lUy:K9Aپ bWǍptޏ.\;.+$Meqk##FdΥsXCPtiF{kd|44&k*TP\dM`F.c҆Hu!QG"p$7c*Nz#xcW{U,]ovK7]/<: &hkbHsj2ZiE2IaMv2*I=!D꣬?2X ӾΪd)U:`oY@GW*aL8&{68;`OFYޖn}㫋ePmS,,vEgYɲToY> ˍLU驹j}TyK)!9 *JDΎ|vlwq`&,PߪwpAjQi (;'KB$Y9Z:tǎ"rF5e„n>\MJMR[_Ao8^6\ {O *r?jSkI(,0 v'N(M+S"٦NVTx"JZ9.1QqõNZjB/Y242bXdڨ5g'gp_S)βԅ[ɚ}Mb s{agSʹrw_h28ڷ#;ja ?+OwN<2Q%(l1/\ͨ-AbTQE #ؘƳx!dv;c+t12X2H`Z \en$?fYQwS jsөř&/*΄6$5u,}csy)Bt2R FXHEF}mfa[mb˘}Q}R~ kҗ qFwW'»4i;O ppJUҞ;M%(¦RtDLyJ^qsGAuӭ''![875JF~\LLk?rq1?Tn~t-*74{Og URyq/OW O5 y[զZD{B'r LwRh/EzF@ڏ&G1 0&b?..`OFQ οYYS.s? o: >ǷRr؞$DۿbgL} Uaj?9!< TBgaP%//<-ΖPzAUFud Ax=Tr^AϠl9O{6O _Oީ4Pn₿PÃ?:6M')»r3i/WsoAW>zB^n88V^<sN&XְQ+(Fsg Y%*x1SPO ain@4U5`w_ \v=Jv!eCZ& YoyAV=A"~ӅE4L K-;.l, :V>CWvon5suk' yju΁-ϲ>b?j&^1*q`,`lni~MgxKz?|> 0Ot\2jOvQ8|#Z)ޥr㬚oY/Tfɳd{|iUf#'O"F8v]6m5mo/^<_LZ[qOu\O/~G~2ЯnbΙ'|{pկ? /) W@ۯ4Hcy4 OF'rzM&]mpBmq{t)e:6蹢ͮ;LՊ?ś?^=/&KQ9/1ya9!\ӍxlKjK#45-, h)MgiDIEO Z+]!`%[CWW۶t(]=AT[DWXժ-thm+@II]}?t%P#9;]Xs~p>/]֐~(l]=Jttuhc!X&Vbm+ H QR+Ƅm+lo ]!\BWVʦܘb;z:tř2p rBT'HWY+lJp"h ]\eZvhhv(]=A[.ZDWʬn{ADh Qnlttt t+agAKn/n:f#!|~,FM<*"p4t̬ )+WiV~gG='~޲hO7O6n}wYFtG|ÀdmWYVmeRS?R7?~{%ͤR8Jb.1G10!v 4x cSH/DB*;Rd!Z|eNX\@l饷ef %ɳSx!R*Ax^;=Jz2#T$%z*Et(>poNx\wڮ;O<nNQKIj)` s2Y&<ƘmRD,)+((`)-ҵs[Ϥ[hjJ Xm<;]X;k?75)k?aNr]i7+̙i ]!\!BWVɦlMGWBW0CL hWȶ6%]=ADj[DWCWWFB PZ;zt%!BQ+kU[ RxBttJQʤj]IV]!\ ўxBButJSm]`ݢv+[]!ZKNWNzte&MkW B(\CɻHWz&c ֞++Z3hu+DitGWO;Wg!X) jZꁏ.&;?K1f:},Zs&}PJҴ jVKiMH ݙ?Zɝ+D4gRVTs\+DuGWO8Ma 0e5tpj ]!Ze#ǮEXq$ae<3VI'Ö,θl\J7U]}H^{k2]`  oG] >eCmD1 %xDWGHW$+8Nj`/pu3fhո@iOcWHWm 0|yp=cZ]@ÉR{Y:m ͍焯+xV*Snp'9;ΚZ s"ٽhN愩9cNlSl{Ж` nf@W(bbeSkvF\it5P=UXu3t5F ]'v(iiQUDz:I^:_'W3˭Fß2 t< /隇hI<~eۚ M+փ̓V h]=] ':B f pEW-۵@yRWGIWƒ†jf3tjVjunt5PÉ~]hU o 1 -v(퉮$4*pBW@<֜+'.ۭ@2򉮎4 od;t5nG]-zjc0qlijjpc ]AM+RNf*F!r8/{_9] zu5Piqҕъ'g,)slr 5`IС1̃;y(׶k\AD/mzkHo ] ;@)t#+g\4vCt5vԕ3l(lZvkr)Ӊ? Čj=yhr_xJѰ!`1ۡ+UBW@bX;] M#OtufF+OPDWCWj!`py3C~j'3xtE^4UuӅkojNypfZ1krm Е2 6{8ng"DV?a?齸o An|x_Q:9{槳^]_Vq햪Og_~ @EP-).U#voP+Igo]k?#]_ߔtu\)2mw}{Mw7 ջ=owE{Y6)Wvާruq3>98s~7]~ŗ:{?}Ǩ={rDg9K?#?폄xvTTT/@$Q~b6{Z&:_P7_WL qwPO!}?/tkMy3_s/ݛd_ؒCIY &j3;ex>$c+ڿ|d?vyߠmuu~qt??@Cܞ_Op9yc"u_#г71ULY(qLax\/PFSf ͙RqΤj7VhʹRV⮦T^h; SMb )d\/ RΗZY[7a"FUl&fn̉Q!v-@KRkPLȁbkr-hH̍,QvrR9EdIbt<7۷ARMR"1/!2viL24(85cz-pO=#zAgF3vk=!1Q7.{PtL\b/9%㵪}oD4!a ZXD*l\IQvڐN:|:Dc; cSo@PB. ^%.ރ&V2{)&XC#,vS-!v8Ǒ֑-~B][F?(K5>R$ښX/ g9i6+Q\"FSy,yrJqMٚn|3uSOieRr`̬Y=T'HT5KE Ȏckm.5SGގ`dMRˌBhn lqb,dP*ώJgǚ[ݸKP\ZFohŽbhl{E}KxٵP@cERZ6f=a=g36APѶ J(AvdZ⤭T.l)S`q_b $؎zokҥ:"RX dlc!}MhHpu ). d> 8׀T %.CiZV2T7 < &$Idp*zbN}zg ZCdܠaS`d 1%: Ȅ ؙ|XS(#χP9[ANEE; }DŽ !.xdWLZJ N u&zMp fV r ֑\_ fU5TTgc(@Hq`j ʋ,coLHl oaH:ٹ+UM.qQr/ J HOTe=#9Bf|Ɛ_Μ:Bj2) `d>"2Db hCSE؄Z}FФi!JF1~C3TLjYcL±)Cڡ;"_0<`Ov{>5TAwz|w !Г ƛAy(KJ_6"LE IW=$B+uXF SayG : V 3A9AJ$rDNk z:rk4XVѲ<0{KH`IQ}&H@ɹuʨx$ۑ< 65pujP4lU;1 ne0 v*"}ߒn7q\2'ҢӢ\t'a9P-tk!*!P@.`UBh̦rqy7\ X%dl4&vdpKN Zz !KkK.?!t=+{5x#7}AAX}5&B]<7?L|i۝G}m?I(Z?O9s{"B\=N.zkonnrPO{я?ίj{wu 'z|ٿNo]zB^|%J^{ӘVx܍;_v{?ozyxnn{S e{<4c?9{ȍb6/ab`d'j %CYan۲]:LJ?aGGXƶE;);GUl#gnzUa0cβU!ޯbSUg}$i""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ":_xHY:dClaVվ4 *5m{$eG""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ":[>/8  TK'@U}'PbHs$@{ D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D t$PvC"$Øn8$ CZUUD ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ":˩k]|ˤN5n _rǜ|9"N& m0,1iHp%T+.J p,Ϫޙ Ȯ`gcW(3=j]U;JvuFv,߉`zTTrS c}Nb0`꽒wᄑMS1LwxäY|n}ܜ=~#omӇ QrM7U~mmzy.1UύvugWñ*WUUkvUUZMvuv%,&i|@vUk5r]yݮJdWghW k `'a0vU`UU(:G]6)cWU6Cֹ]])]Uv8ʅWMOV>*wV%:_wgW(X3r`+T oJdn-Fvkb')LhN`C?&ިVDžG/lx߾j_ϸƲϧ|fy@\s~W `i&b痛^{ʋD<0oU9˜UWk@ݎi庹mr3,Z<>8'X3f%jUee\2Ya~i9^_Y뮽2ԭ:z_~߾Q%C]J&"*"lOe:r|l*%x k5$K3 h*pIv J~J@l+RӍ4ug_RԜTZ[Sj9OMujU0L@j!*tBrCDZ(]u()vUk{?"X0ؙB)Uo@v9r`FZnWU!:GrNҥQ;(w *)ǐ.rw%ܦ̕}c>L}Tʞ4'>9SR acWU.P쪪5vUU:FvuvCP0DWULtUw*!:CuB= \ cW( ŮPOevA]]F l8vUr7jAݮJ dWghW)3+,p}7}顁4vuv0dW(xHo\bWV0w*9ڕ!]`k& V/m4^mnl~F|jl4m 2}zn"^ߴ -Fi7>aB.ƳXr Zn^y{M^m~Xߏiȵ嵟Ooׇ̍gav#<˛cl8MmyS w,ÛaWџ9"8BQ{ze[]/צ~9ݍWjlc?ʻ]`slg0e&h3\"2 Bs3 L+̳0A\vۊ)˩ҽNvy~b|(lu+η~n2eKZ*0n?:=7Wtb׬;\C:`'{\f>0Mk0S\ίtwyxs]+'^;56m@uiI@{lYWOm<_tO~6ĬM<Ibut+em}AlhqQjQ:H{Bi` &׀UH.O qj Sgd*Oό\?X>/NvvTxt\iBDuTpZVEg xQ0iT%HFJ掠G!Y\fS9d+f6!S6t)u&g!&M4ISjf*2c2X>=\Ѽ';Xy_xЦ>V^L]>P>u=Qg7!ܣͮE{QtdIɼ =q>MjY TiL5Nހ(L)]ãH!Km>>wEF}t-˹=2٠0:/]"{33v`hFc FR).TuhĊR>\  `0E`hdGC7O 5{"27-k^svYiDجobZfAg=,Wi\iq'"Ɩ-tu*Lbόncdlug)xߢN!uy΢xnI)f]Xc0< medMJ%0T44UrsœV}dpJ.Kw*h8L3ONGYIϪ̷vwlvJ4~KRǧu~bw5=sox'3`(VqRJ2mpN <ϒ:TB`u+mm* mʤH>'[ xZz&յ3v&g3vt慝cPw셚EpmqW#ˇ%-ݚ4=}y\~/Ohx<8_cGAAr"h0 eŠL` $V]S}aQ ކ$ AHQmSj5  MYW N &ɶW]a:I;)f_\qk]ېkSTؤ1#,8bJY@ 2`;q5iα5B`S?,JFI2F02Y" 5}ufRELc?L~?l|g5|(oLiebb4 P&h$7)F.fb*+U >f ӮȄQJLؼ*R}Bo >/og+_{ ^yMy?t15nq^i}vPz5^136S¾K1֙`u*9 [lG"OxⁱdIu;?2:}qg o91Fp<SG+Bx[S6;Y_A?M /b\ꉼ27,%WhsA f*Wfkپ/~*.~nl-@ä)#U^I{!iDYl,M(:3iLTV"7whq}q7ih_g8~{UZkzF~fm/@;sWsb!d9F!&gI ^dUs_lc< sY0xXvq@K('~xԡLZ )XtmLrSY9W"=S!GxGmyՌl.)w,{l:`H LEJY!Hꎐ: iʃ"uora'&qBaW`  (E}z9_dGR&^;//4_XtQe$: e@g\.!kA( B!?ƏZˣzhg3J|l'Lǣb`y܍S;cǧt};[$?g[l! ۺmvIPআyFHyw3eyl+$4{"#G.rK#0TcϩHx r4ȾMjfLMO?|;Ⱥ 8ߥ>ZɍVeBvyހX5Qn )kۼV~- r_eDHȋ.@MLucYN&X<ޑ+f9(NN=oUOvT^=2 i BQk9`Ŋ:<^O d"-$<:-u7{;dQ-޶ƤU &@W2G㖆~`]Žy_7̊# ч׳ӎ\raWB㵻cN,2F'KRZ%1+BQrCӋ:0D!z:Vl<3Y,BfK hj "EG덱 hlz]̫0%\ao񩧪W/Q2t9M\">;twINw iwMiY@, Dq{Y)0I)U}ރF mKs~W7 y{}C/NHPAEO$s2JaQiƝVc +"XAPVdn z EA~CX;ݯƚjdB[ă92S_l`ĸ`%H 'g^#? WO#vYH1* s*D4%KF|pB#ޥ ]H8wK fbgY8.#P46b'%5Xqdx+yNל ~Yמӷ#5ְpb\Pt08^OWʉVB3JNNyh-4]g qrA<LqѧtOnZ?~n|_?z6j8[7!'aܤr6E /j`%|/]1,<^qyveB!ՒP-1fHw3t6Ų!ဋ?' Q󁮖mmӲU:dW}հVɕ.ZL s`N02fxOpV6__nWЖJ릲a 1'Nw|b_]q>2]wԭe많R?aX;)u`S YmAD +$q^*M62zmͿMd숋Z}֛y#f[cGw3c8GB?}+*Gan? EǸ/[qكcH q0yB][G‰uXyC:ډtB٨U6+jYֲ {X~ys#K[?mvL>5D(O O=cF@X1NWL#,8ZI}v;it,ݜWWapb%) YsƼ E6rId\E>'6'Y6 r7`ۢ|_t%7]eխ3x0/<}ޣ;ꢀ>4$ձ<_{OaނWVCkmUAUgR"͵!K(MVx{OxĘvb%:XaٷP0N/Tj4v:Rcib$OpϾȹ۷7q$

    "Kh8ӂŽ)CVYm4<ЀymR=8 fXdTaKa,t"lH%1QMKc5!SR:k(4)0ӎӆ {F2PJ%'(d%thz+:tt>;P+BOrqCZvdg+Ka2W٤lRg6.оsLr2'&;;EDXG;CsYk ]6`]°}V0#A|ۜ-rm|Dcx:^g[QwλChTm FJ#68R$#Y@SJx->񼻟 ztF;|^kaaC"Hih8?ê܀iy6*U*ܤ9Qln%9`3}Tb1şS^DR s&ȵTYU+/iEa%)GG&:FMsL(QhȹMDg;X+\$JK*HχMs.tH/LF٥+0ڰc~T0W _Qߣ} sX?Q٦S}\<{{,|(#Ilrq?j9z(RÊrA\=Lտv=*Z`ˑ,g\\cF%Qu8Zenӎz2p3 ,`{ҿJ"C% ;W(BB('n(F=jLG eZIDk1hFs+%}E퍜ҲRs!! B2[/UN %Ĝfn9Z i+ϳ2Z(C:IVUj*USx4 L֢8:_N m6]#T劃ƹ:sʘOW^|A ^y ,kPRΉ;_LTU铩Ғ&oM$].$e)+3mZаjmXWk2s*j"G[ eha|TlH~Xe.`yΦkNFD#/d[,=bL'b7xVjwX{C \_봯 OdU/^YC>ۙȭU%,d1@0;8 mŀϴݶv2~E%tdX+"/M)8Ȭ]{˂ ;W0ËWz1h6X>UyΔ8]N-(X\Fc*݊Ounyz1,a 5g"J%%P *:%krRSeYrC=c;kwy=3vJ^c2Y[ƬF$SɁ~2)Ip*(Y d, 3 ' qRO }$W͈}Pڄ0O*k3D4s>Y:!Fk8iu>J-lk\!V)+2,k2he$u2Q.k02&i襏$߼aEJyeW5e۾}{~hUo$eG:9yK>J.s~=su;Z}CwCВ^\Bܝ&FT)#ZHmBlJ&l|7V$bQͷџ/Zu;mw|bPZ Xffw]ulS z@׫H8>MSe D(hiCrO*TWPyT)&M|/#d5!{!!lЁ Rѹ$1j (YDG)ꔢ4lР. ϡ NH"102,9;c%j@$2qLt.3۲m #-<|1|!A^?Oh>٬ޱduw/m?sN`*, eE&"XDDRB;,X|ArE2mp7kgǷ3I͹T AMc{¼ZcQ(lM(BHٗ:P vM u4OeX-*p ZNW35ҕTB&W`_jaNWes@WCW}  `۟p*hAw J:@"7\ `EBW-tyAa+m#*;S.UAGR!ZiϻwM.2yv>t֯Gk#Gͦi>G+Q;g!^v8IvgQ6 gvL1xu;kG(\nCh:X-/;f8MoxQjKSoTN%9q("B J}ehFyQ~Tv8J نv5 5Y5Y%8+b"DcZeUBlcQMpQ} -οo5(B Ld\/tUοo-("zDW("C}+Bk:]J! fzHc=VsFmLMNG5r˸=7`0*ug)1o߽-fgUqA<1jV;W߾o~cF< WXtp?M? KOB/$uM? 蘪Oi9C^ /tEh4]+BiPtut%,zDWX *p ZͻNW] ] QW썺*hu:DR}%GϥXUU_誠Eu"4t67tUZ" N骠3ҕa%Gte ֟gW*˺NW46MJ;a6O nj4HFPo,ؠKNra6q8`SI 0y\d>oy3A$[[%4}!قVv[A 8劣]`7tUބ-v^vv 9pӪI|VV?iQ Μ'tUwSO+_9ih> XO`iXzߡU"UAZE@WHWBt`a+*p5 ڗtUPZ>ҕ^r2;+.Rx!-,` Tv]Ӓ>, 'w՞y[w.hxvsNe2c\q'7OD厞YguACr2G/ߟN7xWq=[?H1&s}v%0ӑ?#&| c #ћ(PI)&2x% aXB=ϲ-*nѨm]z\mjF=\dGrlF\+x_VeWyeiuW=Xh9zW<|wXq<ٔ}6H7\,fʫR4~:}So5Uqs`UUUVx^L)@XXߖ>_^dO'r- YfXfץ7lTrԠV)ߖ=pH6k=CŏޕqW{z>ʤ_wfs]qƝ-_ݱ'߿n׼c75f;CFj+5ִk),լ+G)ÓSJGnm>&gf$SUYjoea(۸']󿜻슃4 eb8yuǬ" ,JJR Ke \g@"b{C)U&WS좌&ék1tuw]ltEfh]4#lkEe~WϧS4Wl5Skv>QkD?̖Ӿr}*Snùnid˓/e뗱!w+ф+:Agnߦ/q8"yAoܯtws\1ݥg5ox~w6m?^OzukxOG ?]44_jrÔM֢l4 t)6$ɤ(_0}9n]7JʋxwQ)l81>Ve?aր,4-gjlFfbшl@_m6LMnúVԪ6 ]2ݟubJ{nN?gD{O%17;\\7H;o*,,+̋9:Q&kN2ӳPr*X #0U]SBtGu}ipȎI[Y)j0ـAh}NBJ%0UhUu)i2mJbZcp[8)\]Jg ݲl êFRvdƱ,C׏F8&rnOGم2Wy@b&'xYS*`69QݣE [uXhXS˿"]:!DHͫaZv<<$KI?-.0_>?ǩ9(]]e{׳kL yFJOק$mz13杮.*)zc|M;{ϖ[Ӿ[? Oov4&J"*E#+}6.`!Бz"%3( Մ FcD2Y+냁 H0<8X ;ANqd:rK7|s ._6O&4ޙYO>{j8M2,xcJK+kYP;":U4 -B^yl=6>i b_YE 624w#hؖ,&lǨdP:g3ߨg}t<%cbϝ͠٦,v,ɲLRlv] vԎlGֵ^Vjr%$[쯩a}^]S)-Ӷe)g_w~ !}kƹgk v0GN瘲. cNj)I؞yJĄ)͏a7Bi- mqa87n'"I<Y+~cחrB}<8Dk!i\ G)ԴQ5<Ɯ8A`Z9£p5g F: ZXP  jA@RncF0)cE*R y `,I K,s$뺾Uglf v8!iI)>@w'I5enaCy=-/U|nysDϾrc.NG3G%`0 R8D ,O7g<^he==8QD F *BTq0~D%^ΓuF ztf^ʐJl%6,}1XΜ7c FiP(&奄3_9{o!˪E UBqH&um fcZ{ّO\w>4䳱; 󯜍?ߟoo,,^ XiQ #t6 p`E6ޜX p +1^guGx4tJ0}F`1e>)<(#RL1y[33al؜W|`PV}nI?s3kX`Sϰ?xl8/fdfToBJiXއsZGYPwKQA({g[\UJxȫ[KVRrԧ*Q`kJ?S⬄`2;WQ]Oه* K"}NB ~ɠzG4#S f?gf1|6hg=Ljno<=?~,zMA|MFƅ}7>;[\,fȒj: eO*|\N'hh8Pvg =pO@A /Qt7/^6;NAqt<(.ng](èZŲ>'h`҃$f̠vT֢â;~] D]15CKʙ|,Sgb1A{rPM].9#F@DƞS r4^/% ܚғ;>2Huq&e^{Cr0FIUhj2ɬ[&jzr{N.5#;хWl5WqVrjsN̉Z$riBJk$3f\(QrvBĉ=۱hEttd+R8P:c$ bY!H}kXkM?% t8b+:Í~Jz][-_1R>Ky.ou5LV3Xl4\JL{һ%| ]) %,]Aét xAHPm0jMj A4Oud49vcxKJû_YK}:8i~kvPdk% G| &qZYH2b6ESa4A`'Sbg8N4ms }gߪ:T]dt7}ifS9W9v9؁EၱkT x S JvH] D4K謔wYЫA[VAYS$A&Hf#DuX{\y,4EжԲ2%oh2d5~>2 uO7F5Yu5f`ǷYe.L,ˠA bR?_fbW<`%ьK2dƧor!rFA`CsKFD§mޖ{^[IC1J"&YNB)Te ۨs0uz) 3b 8Mр_ acڲdΫXxU糧6f' 6\1O@ &Ju3 MBRH&< BZDoml~%mJGKDz S3KZ% my&w ?aPl< eS*G▱?]\pcHp09sH[GʠzC!-.*vV撋% ;:}Ǿ E$*8S:)BSϘ8P=h:FЉRt  NVFߪݷ%&zqyֱ'3p_]~dr)bڢ̯OtiM m|.sKqzY)0)RnCp,i\^ROs@k) jP^G*#RPP" Rf8b#+A9=,?HET{ltBQIsP!J@4""#) (o93(F$)'n iD?"Rݳs;qn]`#@46b'%5XqdxRݜG8{-[|ةrV.//]V6>hɥh%4t1枻S/Zhu/Φ? A'h%Ӄ0)KTΪm_KU˛~p.C˜%0g˰IYSiY{ΊqQ>.OL\Q(Bm3 :7t6 iFiXM.ŸqUe>Iq1hzsx6rZJ6/6kj 嬳c G5j9ϿSN1K}q 2P 6, ''y_.}/o߿9wG{-p ̃K \߇?cj0_35Ul .ߺ^K|yFh-EpI@_-̥&''~(d<xybׅ|*l|8\=Pq@T;Uo>. nӃv'Mgc`)1Lr~mQ˳B/0ThNHӁuk1>?2r(ho fkvwKy.{+}f<(|9~G*Z^~/rGF\PRy2 ^c} WEA5WLjaĥcLc^F%L8x(ie+嬁 &4 F'D0^Gر?ph*u/2GD/Lѕ^V'V?rljա"kxEykB$1] 2"ߝH`SD!m$XY.55Z +՜z8TH ;H]$Вǭ޵ЧX`pdqoM0)3I.))vS=CRԃMš5 + AfUb kef28E,W) &!(Zy9L,h_(`<,JdLk,8*35rV@܄n sJ$QY eYˬ3Y{. KJ^X!,kniCOJCT 2s!=O\P8K>6|+ݿVܿR0WQ^hn ]Fnb7#3܁6,{Ę[' tV9ZrD#Mh+[@8+HX̃v\Pυ,YN=\يR V@pIeɘg3(Lde"N!6SP.JT"9ADF/6י@"YxoGm dV˞LkmW}63[ps\@9( mwsqclHNW]9]ݒ=@YI_4# ]D0@JR[e1 9 ] #ZMJ.%@M)z'b&s#w;+U^/|n$DuJk21 >9hrZya W Rn)>Rv_;clK pƓ/?_ծN&i!< 0Yf5Φaa CbjVF/W&S{_ b@+mȃ`m4ENT<<5BΌһ|ZH?GoEA<53SJ :7ԋ=vޑtJ/'|pMs#MaqDe[MpAEMͧ8s妏gwZא`ɳ!ӧJ(~rWklp w#}3%,Ex 闺bhXl>":x 8p7oev>Z\Oj6||٘D>`&5#nz񗧫_O9=3Ow:~t6.ιu>6|ݙoys0yg7cބI!ov$|֛>݅ʨ9^bG7=3ϹE\ ϓ>.#-PVaRj5[`6;db&ϢuNKT5xw*H>uH̕1P)(Li.1"C&;HeqުP4ʭ媫7/+Z頄 Sde4 NJi'I!r@%-5uH:^#BosoZ%ÛAv xfˮ qy>G+ ^Ud|ꊂ\. LMzHO}u ]SibR\NNhM;OQ32usp]֏_/.kbhoؾol{OP"=yz2=]-#ܭPx7eu!vmj(8}sҜGޜ+t2MŐ.9}OF{'dOJfWӴ 3O?~ 'wFZ%EM. /k9m-r~Y_;'tvKBLd:<)gѽ$tml]LrZ^-8ݻd-jݰH]lTMmw~;.`P.}j:-xGwo4Y_WM&[^ٰ({= nlrNaU27?IڔWfEY2W9 *fMae4t4t.C-pDiV*G̢M`5)/;ș)D'AH2 섍1 }]A`du^HCiKQyk<}0]7۝),2;+?:`dKK^?g:s jBμC:y|ګNL^(\QoҟXRe4~ G{z_p"* **V9ń~Ya #Ռɑ݊٠kHIF:L^[@":HQ &Y ja|& ^F(*Cʒe ȁV "9cd\H2-;aԓ"LǟtȡKVI2lt5ף; bԯa?nLDkXAJQh˂OG-DZ$$0,D WW>HoIG{Tq_Iyn-d cz }^8=ֵyi `J;;(< nGKЎ\bZ˘BNDit٧@ .0v|&9rv8?j!Yz̡CÄ0IN!Ķs 9T`5ҧ:=%:=%zUNʡ$2%D(L %wCN/Y:>FkGqdGkq؆]ӖhԘ7w~LSVܰrm.<Fd@]zJcdMyK}>^!ڮ*)г_k,A·x\Ru&zP54Sڭ3b% s f|\\ӟx'Kglٞ l+jO6AKjw=RՏ,ν;~hڑ( (>@&Imv.|(oP]N'QdFAS pϢ6k^JX/`m j{'jYI{'x&?/ v)]TR!7GN[8tBǂN[՝G-TգӾ@tZ#\YEYwe(y nT|L&5Gs0 }6HXX}qjBreQ60L۰;`d#*ޗ )Ө~|Wt=j08Ly\`7*-2ay~͉ '+o,7 wbYgu{3pYAwݴԂc7iq% .k֝xs#xCnrrZܯY{XM6w:# `uD!v!WMB؅Jۇ/1F{jU]VQW\cQWm2 +,Ы,(}uEƏF]r=uUݷ뺻*Tj7ԆK^3ƁXC<69+Wulr}܌p89?A 0tw߽LǗͣfgUq^q~& }5g`5|zqjW_E{`5usX5U"3Q_GnijZjzۥYŎH]qř8uUq,PUUrv^J#RWDhU!WwUΫBRzJ p.eE]5jLtVؒnon'8;]Mnfx~>j'GS ds:16HSdcuƮxJ牸~- Y E9K}*_7`͉ *;i?F7Q2RF3Md5d sZғWڼ;L\zAQxYĶ C2x.AD` HR!PBm#E&J?͚Ik\>g.kos q&9ڑm5WT^َ7 z+u3l9k˹yGU^d57j&ks?kœ&UU'k+尶UNV;!]";.@wCaĎYht2KPW2e49R0g*[w`p8{hեN(sVc" Cos&吕w$w+ضfij~(Ri,7ޮkusR~i[.'Nל8/$jJFTJb ST`E(i[.1y@5ۂ2$ 5Q$p ΄J!l͚s_ }au_3W)Pk36&5sS&W_}oj~Faej{ dh4 *OVI2(gRfR?!vWϛhdk./uS{Gnqh9ƽ]SFwS&ma|JJ䘅JIVxʫ`brBXR&ZJvI^9=j jdFmv0/#axSR$0}!zN ĊsցTNv@v\E+GkC^tY@={FNmgS;d^`5X $)M*f#ޘ"9FCT"3膪ͪNӇ(}B,jϤP1P/JIPʡxyz \@- &;kX iCtԶly%*'kllvIZ'dVJ Q`ykSV"Ϛ5uٯ~rGʧ?/=vRO~:d"$Q%>\ |8$wO P?_utHEi(AP(b"fW%%kjFp5tҠ'=gW\IJy @:!ȐMf1XO9d  p,|DعFgŶj1\;*Lu'QCiV`k QgIo32'#˖-J-ᶝ/rWCZE/ȾΧ;8J $%m/F;b5'F>; "/s"dͯ,ܣ{\g,d_0_$'i<~#/6\1郍Z.r0P+?DSRrz B㉡򏎖^R|6b{cZy,/ёߝ!Ũ@Y@RvZ _cW4o)ɧGV7X(\ ~?ky|qv>SՑhtq_WG =%:n~rs#W~7o靈UF{ h।IhODʠѮA 7vJJB e[)_M\^KZ=WmC0ۗ\tΠ]Y&ݯ~H ǀIBEdN"x 4A=Smsym0D>/"dE>ht^UHY,;Y08eeGiB$1B`.̨ J xȔ:JT@͚s#tw)VjrweWϽOY~]){Z$w/-Y|}| 77 Mʳ_izcܭ)e1 XgM1],0L;(0"f:N)RІ^)u:YNYVA)@Fa)i(WSҺJLۈ3WKdk'!gbMaL)W@XR *lݱИu6k ه@%~ ևM%b6E;5H.'#\`J!Lʠ)(`9< ƦpA`TIZBm@="i:8Z2;j< w ̟x*w<=~׼c#ۭ|M1i|2;U=>[^|<.Kv.q6a9-KřD,ld.(SсFnW{zQV͹/2ĸ6uP}[YɾbpqŻ,(w$R%Jb!R(D ނ$$&Qg>z]~Cfֱ'J7{4>]m<ؑrcR2+uVApC E?:>1㦌i'q5[˙-6\b?. dMIM aán)Etţ+CyiIH:aCQ,(J9i. 7Ezcz#|1iw۬ߟqkY竳5* 'ϨDP%W|0L:ߥ $/^eUW a m ia`]Re6 mFVp9k86iRHqϡAHq1 Ǜ luw+q/{WR}륯etXϧmD"D:X%+βLJ(AM)`$ZuhY-Z&`B*!#ȑ &R,fZP1dA'BX+gMԱ)0E%r$E`(rNIΩ$dK ,QC5~`&Jc>Kfxpzw뱃6wvS;3]F/p1 0NXN4*'mDU.ZY2‚y􁊱'9O-Vg43 ϗ<*`QIʇڛy+/EyGBN8;i&~hk`K*zAhA(rIZ>'iк9Y%Rپ I+5$wSdGojɵrih]_>"%NO=Eƣ_Rv_ ?01Z):.ۦ!8~5g|2ƫ> h|>NkDhùjB%OF)r=6>$js}WvDaúM`8#$kgevܥ[P5ʊϣ7Ъu\}?<^jo=?_X/Ho=N;8}YԷcZyw~ylqzn<x_I|Q>^-}Tǡ[}/cx}mnO~c|<}d{;3) G0Q!NN {,ogSNLgZtW7Zm/f&ȫV8?/ֱ6m)܇FCAIpG|8e,ѷNÿ_qmq.xryGS̞oe=?OVGTPعC.†h @F9܄7쐓џlo'_8> cjuow 65';6={~__Ge1{?:1lUr\"U{`S1q°<e5c5w kpB NZhS$?~cduN}g:lߊ@=]3uWtNP\6<L;gMط{af]r=Kq]yoɱ*IH}T_ı7ϋ5.d!iHz.Ґ45h,[3=Gu!@Fg$جqCXlvfV&ɬs5-?0Gebܟ7P'l./$^u.Ֆ j::+Ox"!ڊjّ+eW=AQF4 c{H'НGc%v-UU$P*D(2>BH3E |$V hEqڱh2iVwLR; k1@u{M2,6c޺,^2Ul5+juevp3}Fm5ZvE_,tX Q)y$gI(4bc" $Bj|&9뽜N|}Π=yD l؅J)] ӾVZ~P/xzP"X " -5ETC({ˍ#ig[i4 =?yPZkcrtIpEED#פkM'6DB-Je0wItOZ#U!2)F6}:SSSJ'y]?뜁3JUjm\ 2f(Fw]Ek=:BEG5N?˃yVɥ PASI|[U?8*U&\5>yeooE1^'#Xc\dI|F?僨t6rF%%/J$HYnw 8?%:zVaj0ʐ WET}Z]ce+uvXk.e P{Cs׳A}աVUQVkh&y;277Ѵ|aTMuQpp`%Ap[KMTu$<1kC~V6Kݛ8_ܟ`=f.escєLA/bٔQ$G,bJH!)pHL09AyEޡMF鳓'u.mr(/KN PQFGED; De0Jъ >7bsi98f\\}y \9Z`5rn*]S_/Q NwwW}E[/ESr/n\2a[U?3t 3/swjwST z)5ZFHb}&ڂ  Q61Nw)/|Ծؒo>(5,ȼHyX؀' k+CE<"BC̈́RHq[ iY reLEtA!zFΆ?q)E t|3]^V&jhuYNwT_1-'(7(t*7;N SRGT)rRJs"!}AAPcsku%)޺M5K$.gJFTpDsJ3U:Ti 6 W% 9"Zd(@qˏ…[G3.C_}`֍$b&:@'$ 2X Eq~)@^A?1qD&V8F1YZAd*UG`:84E)h4T_F梟^3Ehq `1 7G" e]^)nĊ$bX0+l9n`?ur o.Ϟ\_^-xhi@scQ)lNRJC܃qk.C߂z7>եUl_KO՛W7BvL[b.oQ?/$Pn޵y"h<޿8vp/jIΖ@eW3'f56|qǓ:X m3Y*#[m:jg_ v4XY4˸|>\uN2Nla{go_=8ӅNԃIA4?/~U_Mc{M\5!}M}vemĥ4.Ar8綹K^Zӳ?x0 ] 4ԉ*t*J JSB糍 8 sjȿ~LAW2"d8~sIE{׺w97ܹNm?z7yv=N_RT(PBi3<1h` ~E> +RqhVOZ3f/C)/r^"v\e(YQYh-g)[ X%+[ބGm9!Z@ :Z"ZftM 4נtF'cAzlE#ߊ{~GŞG&B A )ql;mvi}`L7D<;1z;r8Wz[_~d0EKn-ªϓHJLk׃I׋2<@'V|@PN(-!4#A#4ATg4(M$C )t{yu,N$ $ϓ" Jx"[XЉ3'b D,!:< iMQ@tB# Dit«=3Ho|.|Y,Zg1$I*X,Qg,FfgN,`IQkO4m Y5$ iIQ)0eјk:d &Do  T2xz[ fJq3ʹO}R''?oj*d0ID(UPP0Xb::P9]n Lt:]Uߗ-ipAQM#QAXU$j&{E*Fi&ru7%QMnpR9ַ7rz!()SRR#R)quh˂ J%VuHE0^MJtC.Wr3!GC7ۛ$! ~ZFNN~r/ ~SpH B Ġ)-L0@BVz㏶!rqQq?:hS`ƍݠdh0Ck?\'_)Jj('{k !Ps.| J.bHVd`8!fHWE4kod *J"ZG5B1d)z Թ2NC(wsuC}^UX֌;{s/e-_On[g7aؖgeW땷hHs'{{j)!0KY7;^k+ZNwKmaT5EIA;V.zK>+9?YeI<62E">BUcɿ2HH}]G{ 6"N}F%R!)!CFH$@r@.`lk8tr|0oU&\jefjfA g0M;uXT>凓+M"jzڳ3a4M-㟾/ȫEClsj2eTA{b9޿6B=vDi4Bύv&~I &uD%XݶdFR DcrBZRԓO἞Bڐz.%;ĬW逰 OjRz\TknIm7MjW3Tm|uc5v9J$O1 FO1LT-C&(99/1TUX'#R=qr@Iwň+- + Uӧ"j%Eh'-Q/REgX {\ )cq͎oݨF+r !^j&CiBl<0ozޜ *g:e] #5s8ƃ~OLQ /W[UH*3[Ŋ.)np0eRZnu '֥A`&T y(+,EǸ.OGqMhi)ڏ Ƚo_k[Fz}{ P-bճxbiO4B{0JɴEcw}J:ܬv)<^<)U$0Gͻwko/~$7Qq6Ɖᒪ=/z=CPsѿc Jk?QX _ѫm\l.k~+8}w`)brxj>^fwMؽ\7wS{>_Xx V Tpsg\]Of^{a'Ck"o40K Z\)8RЈtԆ2xgPPJn5G~| ^~Ľ42;F#* `SKE!5X[R3JuT:amX\nk#t46꺵{&߲̭ڤ9+X0tn *&öe&Xq@X=;kAR2|%gS4a+XI6> Ƙcxxyk3[1N8d4 Ӕ&7s $IeήNW)DKG&eXviF4K.kYvMj*}sj)((|bR1͜Snkέ͹csfgdE>Bno.).;䐫}Qq.'F %Oԡbdԙ~:Q`PY9U' ?:x#:4p(L VZ&I:}]^·iF=F3DmM U9O]Ģ4 w2&F1QBYO͸)w}0B AUpf JO7Q xpqʶJөuALY >`_ɤnPq`P-vT6LdFC*<&QS52p('Uw7\Gh*FB9"Kdk}QFx0%!_QRf7z5g 'i:EIոq1g{W _+pTntNGg9-}de(e'KU0pި5I-V8nj휭ZL[m\<@Fnaͼ$^'ie+Q iW4e-~mu94 we瘻,:j/oSNyl=6ISͪ]oz;ƹCQHMR.t $ TXK&(*-҅HbpA]&b|{2ւEd!m$XY.55Z +ӜzR2`/zRG/.X1(:%!EXH# OXIa(`"SV)4FƐ6Ap^JMBBX1JF:wۨD 8we]nv2 2cu$"ZzQ4rn76kΊX).®ieLݱ>!U} 9U KF@ %@jqSCb4؄TKKpD@`:#m =pgHḛ3#p <0)K-? v 3-dRbJ)h~Ap@H&̙#RgQTqB4Xʢԫ"j{=b9&r( Qܦg;0"V]ks;r+x4^dM&[onm)qM\%[?pHQ2(Pe[ q87 3>(ԧKS!xU\>Ѿy__8s=ΩF02AZٵe~i ߯b>Q[8ڮGl!&fm&PоItMmQƌƊƈnG-By8ͣͪp? Ays$04rG_Zՙ2z8)Ŵu3kVTG~lR4M>8yko&whU<f}79O꜒~Y'vd#Vu$ja39uFwMOunl]afun}{|H6t?ͦClYa˪w{~C㝷-'|X|z8?a̧y ȧ=>x=U=QHN7MOMYqzr,-ʞ}٩[۹F綻3촽Np!R;mwsWˉF/A,ȁrPA7H+$䤎GvØb p)-Sʜ*>I#"LZvd ^#хG9z(>K) 72 x*# Se`\\Kg,[sd!hiF( >JOQI82!R,pe e3,I+ƙ4DTCN!S%騔nYƈEN9'YWܬC?( xC )e2D,%E|D8x(ڨ'T8V (w} 'ʇā&C #"$x+# |+tjtIOz'EϳD,^R(&eES=˨2_ׁSQ ͹o+#.c+U⋮|ŦȹoIz{ФUh PYځ8o)EguNyJc ƶWafȹ!.&h!Ѧ7]I4^qk:7@"'ld7J6ьjCZ.@b JP\pUIPy0bȢ6Ps)Bs~ҧ2hZGN̳AI6p;7XJHF b̹z70¼ZC^ʆa <lb«IZޏjVR"b5WxeAi\]HP\P~ ˜A!JOvMCMȻV{) &̗`e&PƯsijOu42QǸָDphiJ#AT9'\]>SYlx||x?duZ7Kh#ZV:|Di., 嵴R񿻵EbdV_U[ K筯0 iC @B?@??@,[Y/=5pȃ1CB5'A[dZe# S~/p~176j[ڭ-O1]mݻٓj]]|)sȉ^焪\F <'#i .Rϩ$)Np-&Sc HӐkf"P t$Τ }Q娀;C@E]DŽs[FW mʫZ [3*a>y+u ?UC2>TYU^U::)x`C^0פDȤu^rH:iϴ\40DpJEA&R`Ir13s fYb9kz`+lv}O豎]]ۜ/ &:)W8m V@_))mSI0u2G )t"HA!A0E=˾vh(Xj#1A]bRJv,:U5b8\\&u]~nkkrH$x?X:|#f)@IZ;R*Q$H0檊9+IWN^ *#Q #Eyoy=ٝ,pAyެ!aϢ^FHLeehMΨ Wxs EOLL' Qqڴ;̠w,jpi} s2FO#~TScQ)̳ʎ<ԑ.\7.qhqBK1s9IqK"+),!H^ \qHw0ίF$O_lqxELs-+TA"Rxh (FA\ jNSmBhdžS&̗03Do-p'(FmYbkc,JA΃kƹL\ $?&X ӠS*'4'BC*p8r{7AR`|NQkZ`EFBND $`5(;1SG 1J3;n 6FK%xv~u:+0H4'LD!$$@4A4!@aI%wC~?0EXDo( 跤(-7}']> ׏W[I`ZYs8gЕfոd˛:DS#|yS]($gRd$pSdfi_:y5_,rb?ʒu>Y9h3i$cj t.wt._:< ÓkYsO6YZ >kHvС?е6m \pnpQg=]#1pyp|CRT5iNwmgaN:4iZmöՈ7gx3Ӯޥ:wyB@`;s~ՑێZX;ƴ;{ʈu?yCcl{JcN<w{nq!᧋S5Oiߔ&tM~)eZ?NG[ [YnifR׫xڿ;m[ m|u<Mh)nt?`QK9w#}>nsn&C٠MI.)VǧVHg&mi릋Lm({MϞ;eSv}6ưCsw$[;- Q)y$gI(4?c" $Bju$MrB#pvZl= 2^fPGAwe$AN=g~B^(潋n-aQ :礠`S)agOWUj-g||-_&P>V8[a?iF7Fӻ'պPϋ% > %T>E/#QwArׇG|}L.x}qQ񷕡:Ӗ쥬ͩ'Ib"Tčaܥ,w$pYaD!2!B2 ϗcjoz8=(|foQ"Dt4\!)ip}٦y [)4.8`b9qADwg% ⶿s>{˲ YJT]{ ]oD|qKٲ?vURWB\Rŕ Uj]Qm0րQq]=dvvC<^D^ipTȬӮ0R*f }>)є&dH=: YQ*9*2\~|>ߤbiה)<8БᚓW\_f_?q)'.5{\_ڍnˬ;(q~X& G\ ?kֻGlq׀|\ccFhڈжc:دE|FH##M&G+Zd-4 狰"+Ii1qTk(ffRv! Ȑ EXS)68͌RMARBJp29p3x(D,>%NLD=Hį.)zՖh\T+kp%u\΃I@!e "#7nH j`;6[\3YÝ4-]rp)|rb"N&~"3x}{!uN%r1Nx⑋_}OF4TVxsɒz Ä<*i+^r7#pؓLۙa\:{9F;83^ZF\EpՏO1cyF` d\zw,p?CWf>ۗ0P+.:菸zj2\ Q+O[*nplz}a`s㩔ΖCy>_$wUB$/࿓_~JXg!8g rjPsWWު9+o?\ *sԒ5g`-ۏrņY6u:fFI@E4,u$m;vj?<wfM n/mQ' _7,T2s/,q؂hm VO˦ʛV~Y\#J[x\xqsuyw(׫cA[!rC2k~<?O 1v}zx_o4̼m$"#&G5gtl4pȱfL] (IZQFC6,8rш,9e섶94}Ĉ7Z@p2eL@j*sgժ!xF,†hq5Tq q9WvF#W9O xU|`ưܹ/t`[ha㉡ξ‡Cߺor3>; &>jԊ{L2z6=1 plp5: Z2#!,fF` <\A5Pz*zrB&vKsPq5Tz3RA2#\ :w׹j}NwSWW"Ag5.lp5jt*q q彉nNsWӜp5P+r*q q,ɝi˧??,hq„p܀w}VM'.{s./8EvS2 Sf4q>si6IP~uz8 d6r3g12X#i5c|n`炫6CP7\55cӉIn B%mm:٥?B|7<1wS;Yf1齚^ W\J3WPWPy`oW6Q{һ qO~;=;%VzBі.My^@̽:koy E[^5"kgŽ[ϯ mנwɲ/K:;N},D|yv|]A p̗{ς7W]mslXқ?κϻ޷pQ{=zUiNүۖZ՛7o_ wuDŇPT/*^1{T˖~x 5xueٙk ]EmtQ. ~9dY퍼sdLn{M %xI5O^>뇛$!>d}P{^^cpqt3[Yhymor'*DFnsn&$֠`BɈch SmNI35@BJ9qXRՖlw`Oe#V$֡B˕_hP|C" LDH` l&S$jABVu,>d?Zj jhVU˵1ƙ,Qczs.E>d뷪Eժ\.޾}wK-z)5[CCc`]$ '%-pF1=FICKI^ЙьZOfx.Gԍ^3.i%d|=B=Mf$vZ_ETl\)ؐ>-9qc'Ok":Mrh*7rzHj. ^%.5{H<쥘<mP5"@!Z7DyOhMW˳TڧE64EJڋXnj(ϠdI̱5OB(R>>[%Ui5z$Y6 CIITڦ{H9YCEPP1z**JDpɖM,.5L)-hU)]3!\-"JM (cdC4a[iJ8X $d 0fXpӎ`#< p#Yi!i*31p ts 㶊EV#㼩IkM#R6 ijd]8[%!QI]LIJi#hv#m!jN=!UHU*Ks p1!ɄO W$2V**< VY~y$/. E@x!5<oB[ bc-"*+ϓV14+JX̶ ))e%) dXn#0\֮OF#|^nh9TX] ("F_bE\ pn,zKnCzw/ h/:h ./룫FYuE^!BLFRh0!^n! UL;fCd#հP%lvwU-hs^5f@b- F 'Нn$&eGh`ͱK,$8K$2 ~)#P bAF9\ɂ_p{ Ü>F/v .K45kNῈ@͐eќ4YdJ j@fB yHm@*gUnKo^";" onAH@Hd_kFiI޼%Zwi̕g׸tؕ`quep2.kmRB0nWg C-"tU  ITb0kBKQ͢Qk]Ь!碦^LW`_p"\x>W'v.&XyF2ȅ PҢT(<%#>j l"T'X[hâHY=+NF9E 6H0r#S#c!*\#"(mͤX@0SA"0ՔqOk[H?9(Ws΋6|›Z TfQUq di»2ǔpHAJf:c+hBl~S#%pX?xxR,6 &)]#9,7i0ݿf2i8C7e<zt'Φ'Ǩ?5նN.f9'F\·ݶdV3|\Ĉn˷+a4Du=~*Zh!ed9%~7WUpδq=R7JK(Z] PGJ tEJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%W ޲>)ҁ)WL,#L@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H UYrr#D@ڽQ>!jJ t@Q %R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)7#ȷ(D;x%oH t@ %%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R(>Zog*ڊ/mi|{}\P;}4n.l'Rl|oK g}.- JI¥{!\0Rr+),Y6O\k(-'tRG#j ]Z7D6:]5HWZ]5\kBW WNW@)g{H]ŘPWZ+J牮!]YL^̆ͻ}0 'o6x6i3c>^j<21`2և6+d՛< Ǔlǫқ~ꦨ_*nț@uQ}pz9eRЧɨ![BtT&}:d^?8;a̮g]5mvתk LW%pOwu0ixvOEhX.-toD'x>[S]EtJyUAdnum:*pݝF]bQu*xۢcηN$u> źu-I[UTb:zEhoQ>.lMgh4[[[(w?S=+6?S- }JG<+]K]5_uVtЊ&P*At+~ŮG7R>w`)#ܵWGfO/j8n'Z09O:8f y5ig[BڻOy*}4}= M_ ߀_9Mk]5w>gw=C{HW O;{CW 폮3tPnm#?t`e{CW 6}vs^d>ҕ2 #j{ 6IZJ'!]ic]5?tc7t vtPJNtuX]cGjpUo誡5]5]CB\ڇ.m"]Î|TUC|{}7[_Y' t>ڵS}CI(eoP혪-ɂS9ѷ[s'Qe?lD,bϳ0KoVogT/y2TKNjV:kWIּ8F[k -~>e}\b4qm+~~IɳfBWbQwO%96'=Mw}T3~3zne-Sfgw:kWY;&fy,Hn*4 bt%p+le?GtS1<.Q<]Zvھ>ެmHKA/u-ke8!w"o>hxGp~C/sQβ61ZyZ}EP`Z 0vFfjeˬ̘c.qnP0e\b4XU̿0ҳ <*D|ڈlrhǤĚK/(R(UN̿ed_5]e N0U)J #UExnMeJYp9jX%D1N4;z9??ۇF''Bc@E!sE@\TT9E6ʓ2pYXR1ze7 #s& YNoFjySy~nէ:in''dzӳYx6+݅"Woo:yoO˿/Ao@MEI'v>Χ?£ٛ`c+su4|fQ/l%ރdE#oxݸ2~Uq;KR%*F.f8b'sG Vlb2[5tSt8rO*奮*ԝuvyӒ.˫i.ry9|>z0v1?VCC6bxqygo`zg<'߾z+.ܫճo9 ?n[y5*孊N}h>kl+Y~8|8c[Do4ONFeyS;adn8P8'a~V#>J-CY, A[[q/,iִ>tt)F߃$ћt\l _DM/XXwB7nrkjQqUX U bu޺/?xƺDD/y/Gf<Nl+u_pg&r۩t2^EE%*&Z--4V* R*[ZϮ W\S֛] ıRl}YCX.4N~~G0G|/oóK~|62uWيlT$*dR (F%oN( <6^evqvK^wbo6ބttyMk9gq3v%=gEg/l}淳OCmqf۫|oIյoj*qC*ϻ v[ #U~ pg^v]?VeBGɼH梲-*;`@= VnZPJ-&U v!vukjl.@,r&TN!jX"v1DՉG/bb2o9dެsla#+H!bÜ(X`Y5Cf1FӤ6F*MHD1&֐/ m0I$􈘊嬫-.͖?p Sy^tj-nʒyaA_abKIķk4{y;A R2]njl 2t"88r s<ר{&0y9^W՜]M3kD B0Az%T*2ǮK14ɂ)t9) F>gl ##RA] %7ѺZ5[N4V"XbIU%0(䰔d^uҀL̢%dbqUg(痳C7JV2JM'Iߠ*[2_syݜzop3~kءPΓ/#x @?h^ܭE|[3f- `5,S)9اwW)b`צ}G0P˕+f>0P.$Mt7*-ݯ_.G S#Ews7Vum L{\816[<꧕c+WBџJI+Ֆ5 ]! ;RT$ר СLt} H-x%9qq>r zsHe".qBi囻0P !ҽ`+SJ:m[)H*!g)KS6 BIH鐍BhLVٜ= 2c*^֔EFȡ-Sp%Ȝch1k}^)2x)]4R8-КrjC6FeTx[Z0"s_nnObJ XPhR%a|6D4;!x&ْ^*Ryf=v?zvn@ya$x\RI5Y4 Uc)KQB JЎB gP+^PW0;'U `Ckc"xXϚ-@=f\T|,\F$r16)D̊R=-IhHAI~CD AZ1ɿ{/ vIKT4E ɚ(,Ih̀ѕ7dI J,"&GjfW4Ior7fftRjdHV=XIʡIhHfMfX 1Tٷ9652q`Cْys'{:Ыx?>df°b '_*LImvV $e˒GQGbcIھMrWO==k&"U>h|,86*)^MKN^r$-u_Zc Ɩ{Ǡ$a^I;>U6tW6ǫy<,;>s$j}APlN&ּP>S7$f۫Ҡ0œ321\PhAw^A0C nRSz?=]72P3{xՉq1oJHN2ae-ًt6y޿jARk(\@DyP%p(U,0W")tGNR莮`Q; \RExۃ$1pY$b!d_(Tj[KP$bB(4'"҆ $%2Y] -瀮]>&vg}̮o!Jy]= 9𕲕_@VF/ԱYD]DC]ɔ19rD!? &kYSY)0jKa`mΤ]b($:9|5Bw^N goy>_e­{7~Ym̩ey/-Q.%&%  (U*oIQ$/R2iiEdϔMjJh!n1ǒKFN+Ê5AD4)iUȐZl9xߟ|\cSV=^Qm!ux {ӓVOcFYYz0e]w5Fjў1K6$G_\嬨P|"W*q)sD u8'HR0R0(Ȑ5$톌)9)Qʰ T%}͖_7paydk8dfb~vb,ɳtIVY)^8FlPls]"t.],6 7d@?iُ|rģU' lha`]bT6:ٜRe](6ʼs&xDI@ZI'4>_y^a)~&Ͻ~\ єvv?npW[w?D\N7eSfyZ2ny^#?n<~ IVMjoߴO@_=Z?-GqMz4..b 0w_ǀ=&ħƽioeЙ􂚖n:ni7I9BnuD-bfzݍw6kltˇEXm7B,qw0_k(Öqum4?p( io;8P^·MhfDdM_PKыX/# TǭLv??ʹ%Cn{eeq~~VߒkyML5|madν9IG\ͣOޝ.?qϯ9\KC,hu%ś+eZ] ~,ftMwuG]a[ }~|. =v]NpY+ ͮ+4'WZDۜFAF#$; i H4M >td8+-, ^d@clx.ZZd͡GCץb )Wm+ ,Sr6`Pzo6v%z<cl*^:ةB燚_YUi9֧DND;?Z_?oӃjA:y2rUuʸt_%_D.O;~>%\OSz ̻fi?JO-]RO٨œX(3B2 ݄s?sX.X~,8,K/D ]ۯ]N/=MP 6h]զӯdyv˫;jv6}f}~,W*Zn$&tugw{{l2Ρ/Q9¦r@޶Pnuh4YkM+obN$EϺ9.^*9͚7R%wS[2}VK`DAʭFV{P!wUA`HV0KHeEぎ3SY[dvtװnzb׵t #]K]NbTDl#=h"LZ8Xz`cb,zaR^SmVF4<325oC8՘mk77ڻhҭ$ѓ!3d6[^J88\ullN$^V@{* ;xe>|At ^Me[LZƷ5Z lBrma▎P?;ۆWglݲsƵ d"3&uryu0[I 2·d(XL1Υ|&V9'e:t l j?}gTut]608gcxB?ܳEߦ9VS# ?~{l?sMŰ[T}jrǬB8Uq*=|!)if0D%#*MRQ4ACL-1l+x #Hw͒\0-ꔜ"0V ,Lc&[Ѷl$صBT:J>kiFk ғ3`95V).Z79nha=od4{=E諣IzH,(TΑ ֊o<J2uN3#puθRBN'}^mC*Wf="|᫷|2N]rXk1(vV!W c%`CQ citI#CG=Eo]C[\BO.:]6u.giT+%6&`r&p`,ldB30 ƒyOs--gpo.IvqW_4'vIVר-zH-9E@i@)=:3V#%3v TF; d4)|&ǔIvCou %ibMbBFǮԶSێ-F|!Mi+ :Ȁ3Il i NNG2faFcJ!e(5yQ%K针hp$v"Sl<ÅSN`,]&8`؂V] 2+/72fq9w)GK$ tnݖp7N5JY |jK.sL2d$#=jȟy$@\9(ٕ~`.#%lZv*[։>̥?KHAl0պkµQ4rq.Fǎ<1,W 쉌,9*+u\Q,7F?)сPp|{y|JMO-r#aGj>Sҡ0"\-˵X H-[:XR#NW35YW$kWYõ:Q:HuuR^Jk* +kE-b^+Rh]"U"\1U+l-b JLh4dpމ0VVW,WW ZK,2dpeVU+,U=kW,d^WқW+4UwExɼH@ZFs,tLZa2I[SN:/ZA5{k._0Vn: po?:׫?/@¤ht;Gm*xi۫;ՆS;v0w,'AoQ79h1y-"'5 `[R ,V .RK-xvzpEr\ZUˆoW]΄pB`D Ҝ -(eR}=-cGq1K}mR"z`Z޵DYѾ~W}\ZUW+^ꚬ+ T+kt-bޕ+R@:A\)ME"Njprg+s,:A\q]{YH0HQ X.Zpjї+VFqY'* vv K4b MXHzpeP $_WRÈSĕ,y~P}ek</:;t$VhhRѭ7oLOm"hmw'E㠦!|@gU{+QVl/gәUVմfAgՙw+R*WK;t2EpvPXgzZpjxoPW3^/uNC'xfUc8ٶC'=m8i1O1K.#[#Si,L#wziW,jpErͱ+SP:X1W+ K\cVq*K3q"RbU v\\#զt\Ѻ:I\i'#ԃ+U-bV+V:U+tgZk=x)qg2ZW+ `-0@(OTxx̐E 20)-k$;8'~N\;+e|MK-$آwa⻐Z'd R).5 9gjVIG5ʍ#NWpb5Aجo]J)`7+|fu@UYl}@\(83N[2UyvVGX~~V\'S~*uaq]/:J*#T+ \ZJ4v T+iU\ZU:1q4hWH0}ͮ\:U:Hr JkZT+,W,WW Z[uuBmQBEb{GZK~ ʠ0P 6X Hլ]Z(~U*qu ݆Rb|n Sq8\b5"kzY:4ǙW'+ [zaqI.(U H:֑^*Ԉĕq`W)#$dւ+RGIH)sӯaen /3eyf.k0K^FX,_aXKw=3kb^V㻐\'߅Y%).|JQMT#aJwz\yBAE"RrU5aB*&>홐RIe"B )8BXXA/m;ף:GGOO=^/RiҲ#wz"\U'טZpj)WKK5Ep[HY XW:X#NWZ"\`}5҂ZpEj5b"F:E\`j 6ZW3HjW2ymňĕȚA#\\к\.!KQT)gzؑ_K/0dl';}=-IIv0}I,-!-"#Ήbbe0:-3u2d nG>%DMG=-`YBPTC譪y4[սӁkQT|}{hbTؕUS# !5l8a"JAH'0qIJ f| A+;xɮv KL\̦2^|#suev>|Yn۝Uȹ:9|c̗ōjDTȂU;UTZHEK$}}<^3osK󈧣b@J^`H2Ғ˓9ݷ*#7A3l)RjN+Kx.2TP+tJ[ W: Uaq"#)>$epA&`|jȅ]WчQvtF%,gD0TS &W\h*peU:"p?p8 +K.B*pe,vbvppA !XL, DdÕwpH_6eS2tA;&TP+WJ W/Wg+CGI [7+C"*•QO@\9:#ZAPIpp%Lf6PchoF͑O7RianGJKhQqP퓥ؘ q U8*iS:jtN-:C-'.Jc}]@*RzeƄ TRNXaT"pp2~]Bpe;F.K&MRw}j7>VPruJŃ}|F΃0rKYF%̫03L?UT"M W\S+KmQI=+&5aGB%P0h$+K-,3\!\ MTRG`MH2peɥ$Բ-}+[gEeԣ̴gcF`sIT7mXUK.TVbVRd{hU R:0kN%'Xjb+%TgJ='0ػ $vF(Z!vWT^`sW$z@͓#C(GTLCE'|Sd V Lr)vG-tBpe @rU2peє0*e85L ]Yrw|fH-bpelK j Ȓ+C3<\)EI}++z̦|~Q2^x8}3./\rThNhο s{:+Eq7orۉROg|!#YM<1'GoJ/&ڨިMGWìMەWEL/Noz#9J ./|WԳ&dkI%VAY|3FKw\ #m۟W-ϋYiuek;?VǗW%ʮmg3̦fv g=>HE)\]w}5swƚ\3C囁Ms ؘ|G\M&FGތͷp#B7v72zuqav-# hmO)DrQ~ r!o)2&:% -a\;\OϧCI^4y[;;뾿] M'WƁXa9/_/~jxPGۯNEqY.Ni+uӔtXZ@UlTqޯm#;K0JR"_N Sum 8e,Okw'1_d'_ece}2-mlGBeBq5~$>://9 xa-o<٭/CΘ`lްpp?  !jopD(Wk+ o[JθSw̲2$=0+ؓ @N'k0ōBUxՔs$Zǵ[ fP)(yQLzK$}]r426mvtЩy$Le~^S1&6%gri>Nvy rҒ܆#D$ 6-f|1~hP\̿Ƀgq,?v9JDe҄e ,=Ysz n;# Mѧ_/A0'm!A7R#`Eͨ~ yJ0z#`}jV`wo) JVمS#qvF4ʺ0{QΦrHjx:[͠\`uZ--j!(DNżwAf|:1x[ͪaPdP`lN=PքT2`Y--Q{H){\6}@iBO]ƒd\r*iTIjǧRQ`f!oidelUB1܅EyS<g딡H1)jjG)-`Z--)7?v4~F}ek#%N +6^aq|'1&ٶ X{yQr0[^}mR(4iX.']_RѶ ;jh)tDVVVm$e=+R~}e%NZl9c>*((tWDlXŹ^aJȺH2epZ"RV3(q#+fyRU6qNZ]ӤWIɉ@%HԔ1NrJBhS`aezɹM򿚘@G%PJ=2ƽCJXl`YƵ[kK=5 i~F!:Pb^A\AK o܋=O VىU_߮.k>mb{'t::E<<0w9og_Kd&RZ&M/h eψ)x7>z9M^;Kt1߼+@ uED%ltv(˖rZe̡n H]r*lkjGҚ~ޥF(np@(@Rk8㲮+j6nYIQ#mh)6XZ [ab?ԕ̕wHWk~3M')ڸz0ؖV{j9Y[&Z ozwT:މSMفm%^y%.}ΐ;7Rzj{pkctt>khP$Ⳃg[/MԆ.LJgq&蜆 Y*8gT 4㳆Zymv23~mW7jj$UC ' [D)dC;Rk& \j}}S}\v{⾬{w۵hOV(W;00&]ck Gq?az*x*mAjf5M-DF̹#6RwqQS |rm7>w,w0$suHƢJœ c%3=i{zOEr;I˜/UHH϶R֎*7h Z:jZ'VNNvٜnqW$}Mk2>TSԔd|޳6W~ 0f. |MvӖG%?f+$%KVKleNjθ%7XbUϺ'ikYeCeRQwwUڿ8nu/r ކ},ކNZO*o}?nPnC\ <{:=Y9iKŕFS HMY ir.s;)gʜsfN9yOW{ NW??Il&L<p)׫5<ٴo2s`*å?IQ {^>iV~^#!N0B|^| E;=`?çFІO(?-?#zDLD |JPOwdMg<ߵގF$K-jPH< L kqһϦ\wJ#@bLTջN4!{Mz:{&6kQ]Wҿ~\O~#Z/!I:L}vuC ] vHg9ט؋Ȋıif7sK~^ϳ1 n6"@_Me9};j4'x6BnOkqIve2I1ݙ{fF&aO:!o?d6G糒]lkY,}r>k^rBgyYBLo^Bծf* ?(t\36yGN?U7WylNNr<"rT+&Ӄ<\mPq~)h]`.#@a7D&ypuR֊[BkN2C(TEdLt{[k҆iOBdʁR H-Xf̗<3xp >/=dR~omlXi|6sY_Cƿ4w*`~hZX"$OFgm.ذ!vDP^zE1 Bv*]&΁~4ڔh.l0ړQ{6ҟwIEy42] N]ϧ*|T&UQ_^JmIB-|[ dS(̘QׁeԮR`U~^}=߯BڱyBzT}=:"j7^.zߟ^yܡY륹cj9j9zZ%㍱ud )Wd*SXy<V#v(]` J w,8EQ^_qڵo+*~X{1uLm^TuH'4z+?#Uy/UJ{t'SHJ.o(pfjlڬxyq^s\MxvDIIg1uloCE4NB-Qo@hn@϶}jC7ueT[?HVNc89AA NPڳ_FvVhF((¦%r]-4ϥJ)tY^NbDVS״cX%`fgX1J>"`IhZH7<]&>ukLEJS e[vIG! )Miѣp>Q-挶劺m8^Ro2 WH=hc$sIqw+~/ctj)B74x0&;gʕ m4QS6228?(m鞧qx_2Fhjh;T}scU~NŠǒw`bR^7qG(%TG;1H#z0 mdJؑf'7` IoQZ(!ms׋7dgSwuPj) vyťiJv'0-)X陭 $mEZ82s;,FӾ&HBm1f:W1^@K T l$41?>\{a@B8*5e0@Tq~ࡒS #Єe 8ɗ%J1+O!$l~15hbcߊ}z~Nzs☆7rzS qf2JaKmi!B+]m/05y>9,k>ɦU/}+aCAYPlU\HvH㖥 D`l3Q-pf4z7B$ҥyR7Ҳ$mx[!ݑi}¥5Wͧ44"Ֆ5 xjY|+YXkik0OVHd:C6P":e(-o HRK+S<0$ 1>fƛ} Y0&#値%$`b3URHrt9&Y6ny%s$MQx(OAO3oߓ4 z%S1u{hDL=cP=C͹3)䜊! /ɃϠu8V$tx\p%P>f 4=EGZ,_^'Hj)=|'>h=swBM(IU] _ 7q: w]Y4aDR[ h0_DdkZHtNNjGA:T?<Ù ;!h@w$2"ԛ\q-rﮌZ@M\HV>)n2ߚanlT%"h F A Mͯ6֪v.oqIґMSω䢭1D\ Y.J t/qQ",u/>`Bk=$ێVB Ӂڦ"sZ)"p7a C_.Mh!3FtkLm V\( ]-06:AOLipjgxd/Z87F:|qzMY^jp9ZUEI_;Lmם+EʼnF>\srA Xs>_j/CF:'rN1!}@\!$vԎK"&d%"ԧrʬrWpҕIKbI& ~YIIx/i \)0BK((TDy-h[k) Lt蛬/4I&|UK2)mV*\.A;0V<1˝&:"-: s_0m)R3&14K[SbDbdI+>[VC`-@繱B n.&_1*uhlOLuE*5Xgu{ZSK㮒 'mjہm,dvbg)בRlL'}D6×FWQud mI1h$یю/q.76UVXp1f$ 5hN"OՖ<'wɪc'Kjo`)^%BѶbm*.-KzS4垌||?| BȻ('OˮA)Hےf#أ HQX& 3inj\뇐ai:&dm`la2/|-Ty @NtJ *ievp+Mn(ZHEڂAqi6>X({zێu澆Y {P5ϒW ZJkGX"qӝi NԙԂBi]P.ǾX&({~r5F %ifQFxzҌÚ0XyιH~ϣNaa+-NԑOP ʍמ cF9Ys4g듋,qrq'F 6^Gv(D]:qz Rd^P&A`6"C,8 {3Eі5[;P? HF& obi@G\N:uItC90) a3,7[_ux,& Iz30֙${Ay 8['8l6tNMVH#^+:a/(2*+ cTJ)c&Y]m%6Ҷ܁PFtSVGiHS_#/螌nL^6wO2/BGP ^R\kQ#Ӿ+j$l~J48ǽ $,́Aq,`M&yҎi'zmvJ+K׶ o(:c^ -Ȅi nr,Wh74i)5e~S#lewFQR3fr\8X#r|&  ٌKyڲƣUd/^(2 0zF&n,v8j=ʓ%J 9,@"'Ғ,^%Ip 㗳nb6f8 2 WaLdfȼ5hko} VK4Yp[IJHt,B<07H7O0qG-qsfeP͵bY4&RM#b"SmxݍAzFwa4YYe}QKVȀ=r%<5ˋyEZ.)'#͝qK OeCX`f@<- ײTߗlK)Ɏ\R<$x;S&wrbӖu?ա%)E\>I{'a!ݛMr.1k> NUwLY.c;S   ~5 nUӺLBՈQB<ŅQR^i6?ƭQ7;K<&7xx}Z_&;r6Z9h9)Lv$MSTE@ ӝ@*a׃Sb _SܠQ %FL I.?~XU~Dʑ3VɬMz|B0SHoi%iIux@1N(CAdH4OSZo鄱G-ac{v{Ɔº}AEy *T#146Q)1X#y/9|q}pL5+w&7ʣũ.[α!pz8VmMg`BSZ<¸%&N`BXpD} 43Q#"Fy{7[:IwNφFiLo!e0" 4fMXxWaW}*E:W^uƺVF*yè¸N*p~HOрj9ր1i|nv;¢CkGt`Oݮڷ:GN}M@u8P,Np,-a:+N |urD[Q3,X0(cL ԁt3MHGFOM P9`yQ5 TAcsqPr!n8lQZ!5M7d<[.V@=L=Ó[1kp? vZS7[@ %`LOXrP}^a t1Ll&ʓS5mdƹ/* zR0O(o'} T oLȝ%y7X)/jיo4t%CX dVX^otXA'w_2f5 "2cubM wFF̑? [~pJ/厰Щ#PI:$}OW\RҰfOO1vy@K#ޅ&ͦ᭏tN`)vň? z0rͻZx'oEd+oUأvQi0DKۿ.˻|+%&o߅>K~U cfIiSsISIe4'3O ;m7n6q$wߕljҩJ;lBNBۓlf[Pߓ?wٳwwJ<))L4~^eS,;x}6ߙ֫nݻn|[̓Þ+䀪0T-p,jX͋H7,#n맘E  zQĢ/V{H>*N[D6-Y6[֦/d[ټTcӖKkHj=7apNPuKƼ?z~wj1[䚣EĘt0:&uΠXqUۗ})!Wn;FW 8ք5w F`L_,r\9"%2<5 O+JKZQ jr=&(hbc ڏq!:mbk9k;:RMk8@!sNrڞ8oJ~x`;0g"׆}F,ז%i*19#DJA݀{iRVHbpek}TK=- 0**U$2UaK~e vZ`N/_Κs3ARLpU GUeIbA*jBF!>i9l=hQkfv+19?%@1#`o] 3,M"sF27qH(!X7,t+6tB6M R3I:4abITM@HX')C-mwdB:VO)QސC&I5"%6B4Μ*G‚ITG|LZ#9Әhu- )3N$Ȭ -9{ ȜL4eNлi^*1%b2pcTWADžqRKpX306N{:R4W)Mv1 *h|ɴ(..1^Kc_R*E6e1 )꒘H#:Y;.w.|MD˅uB[ ~^ 9qkXWeo:vbͬZϝ]SXvz<:!4V.z9yscSe~iF)B? ` bUǯGi`=Ii)f~60Nc'>W'V)?kMeT³ZX᫨PD9˝~?>q[ArZ2/b ʰGzH]}~b4Kx[nZPSB31D4#1n9eY1RBW_5!4nMGNU#8AFY55wn6( SU݋ 9 ~zUKW3VuF>.ݴq<0&s 7i[ĨKH`qaV|6lilr/̖ȍW1j匡Hj%1& a%#iH*",yrՃA4z Hw$R3;jtAғ4J͓>kGW3m۟PԇHD[dnWߗo2rxآTQ7$'5D7IĺrByɇ:X\og:Xc"W5mHF-L-*;qEhj%L-YfG@&Qt Mos<%S/=n:SR,6#oNJem͑ #x]`(Pկ~^es@ 7s d 3sq='*P0Y^mVC vZ];Z~79^}kт69qHprEҠ2Z֖ e=5n,_C)w_IPT A1#' IJ[ mo*k`pIκjQ5 ^AcZZ>gV$ \qL*Xjv=  HW4ϧD+4HtC 'F&fhVf=WT j!?Uq51 P* `֐+DuĻMf^g(%VԓԔ_CbZ-V)g.T+s{CaONB{ QWU IP칂Ƨ?{VzZǫ1єh=@^ 6b$̪Ak-2ͭSVD. t JabPV %MϞJL8]lsy &ѴhX~j}ȴ({i=iy DdMmJJ\g~o<+?DޥnarM4ٙb0h*j+o2`E.6Q/ JWo'_uD8E[̤.oaT i$V? e'ۗ멙woXשGb >I',d"?BV.l?~i˥B|_ZhǷ[Dy8kDh+5)N$)\j(bs'iطMI (xAnvS0o*2Mhnyx#"P1,ޜ4c1 w˘2,R%S/h}߫4JT^^c_z36vtJG$wKeV$&cDjcX+(fX$2sXIl8aw=K`#)c˻ .S@([D+<(j\ 4탲 z.=8L0,#B MFu `,l6Ɖ?QGI"%6GTGKݬ̡?J^ |7;*/ =7G t 6+ep fL5 Mp=jgg1`]X&օkh!Q'<:߫h pmzM8d@c) FBX*Ttd0&$iWK&G.ee(Q|i; T/C^!]'.%>[ТI >$]+[۲x?vJPV-,pQ*].M*]6lA ]ş `LeyFC N^]/2&Pxz 2")ye As&,ylGD;oz6!.rq{﷙L=I7}1ڛ3D z e`@hB>47t7o1c˕q7-{\I')sfKAs(&X㼣*%~ٌf;"1#q{ ~B#w`™<ĩ1T '~OKBV-GykvAb.@z-H=H1Nj)y0}< /X ð 1Ώ+QIuphDX!Ez={939E=Vw{4b.XNb6Hm ~HtzHVoDsUS r:AAGLM;A]1'xa/ʝ7\k]Q-"hΤrx9I!o?COݚx⟗6s<4bMH9".zj֍Z?ǗJZDԡgePNH nT$ 6\/7bU\Lh=.r;H S^Vj+i[IM OEE给)?H^oDH9Ȳ eFTT0 !p 4c¼BZc *0[* J&Տ@x2 evgisH8"}ޝYpdѺ5g`T`DKb,6|k?b=O"HNTpa)r1mmnj )\\QHHvCVRʘԂGLS)Rp"4j҃+A1S.Yu]#'0).6R-.KXI$Q*lf멋Q dVh50ϓ($o S!n8Xfm'J̀{cw !Dc[1:gcQƀy}{,]ϒ7v[PciK됟%R 9 'P:!#\] yr._UGJWw1 (qط>L8`:8!.M-Y%JT 蕚 w>RDfa&OvʫR 0.g>< TȲ;;')Kz"i?A&7)'!gtdm &*3qD9]F)Rh9W"14{d䩳8ZM_;b^9'0SBɫ4[{!ÓO˅LVnwMʝIH=޲2 ۷%.qoʚ@񲴡:$~ZHPZt>q>q󯪄N+"(GYnDPdHxwq\db!Bt>2(q[LsJ1hZY̢"M0[=3ቔ-4=DHr+s1X]؅_Gb1 ,`:[ |:z[Ya~0Z,Wqh "K>pJI]Mã--ʫF`mXm1>~?˲ֻQxoNj ?\;MaaUc yr oJSyݫR/LXm sWnA$/hz@L zC$ /OA('~ k `,Y٠'a^gbؗ^Ε#kW` 52%Ǵ{j(0'"0~e xbwpGlT.-pzp%e0ͻH0=#JD“ʬ=x ־MQ #H,'ogu7+Ø7 'QmPd* qם.Q)V#@}TJ-zyQDN}ڿfwwU_utK@hv.[(Pf,q:P?E'HA3Ʈ ⹄N{󋑻tF@"N):}] T5z ?CYS3[FЏba=n=]0oU> Qz 8'4 ~ 6 6la!T w$`^amyAݷj06. gM(M`%QWVNT1Dp!B̌tM-BŜ^R"-0Wk)kobs$g;=V=KUc,RXew^^eDqܰPŹX0IK)N@m_WuƎ9ӎߡSʘ$5;Zk g`DpT`oͮa_+XBd|̩ ؔk- ]a듵+\=y- 4'U^=Uw2a"8[-‰!$oj?8B=xobbRCBxSFwF%wP2&8o7;?@9l^D-vLN:W݀̑pX1jF@xAI@ }g@ iy[@yAgJ䋨Z 6MqH!3 x͍?xF"1 (HUxM SwpUq6'pύfR>OuqXO]!ŹfV=c+LB4Ќ˭q+:d Í˹u2B,Vo R!~OV#\O _7z_Nƶx }̤_7UqǪcU*jU |< 'Q :2ORZAAT"M&cH!ƎJ|Zr?> ~ ~\4=>HaLr@! sxn\e?/0AV~/FK~<"^ķLۘ FIg BI?cIL(}ޗkv'MZ=t"$$V9K|>|>H⩱.28P\8YD.h.ܤZŬ[j}~t8d,wm2z-d%Q%E ;) *qY}A)pJ) ͼZKJT5 "}) ! bD9wM'9q   42H}IHŁ, ƆۖfY ,,\T ܃\W1 Ta~J.̢g*wd~􍜴(Cn/.?+rHƐO`R,1C EZX@ X\Ān^<_ij]`WBO bۜ)w+O!1ǾwEbɲݪڂbA!S| d6CiI=~QNc>yˌ4LgB0@bArdjTBcM'u*>X]>ЫBU5ү5|0*VSQixFms%fyBP'6|X-d@EN/u-2;)j\0y \<3&a;3%Fpa)D ip)+//>T=.}<1-ĭ7" ;T5K2_-sєm'=4[v*P@{=+r9d EZLDi@ƘkQ4Cc)(@L\m%B$B'+Tά2dp?nf]T![vϧq^υ)1b28ň;@jr-΄Iku497) CC9ޝ˯;Į'ƶbIjbu5ȞmFcfe.غ0X iDwj"!?ޙr(ZͲi1BcO<` E;dW? 7wZܸP'Tʽ%#:QTZs_UzWEDa Vb=*zo\,ݼqQjܡ㿬ܝ˙p-L3|~"%y ˱VBhF M,Oʾfõ MhJ0Mof0-qTKmN8r ǒq2dX&Z\eyym?ݠ(D ZHڻh]Tz V;}%#$Hy"T>Z/$QBe6 ;"ge MYB$5p6_4-zZ5wOl $ʈrF`ODz}A $ (D#P\&ilUrqڟ ?5‰$D{! a=ݤܹ)C" 뇴 L[Hb Y4DW [nqt:ZxLUtI[ˢep&WCcHfʹ 0G-נ8I XcZxS66jc4xBTLh٫kʦ򴍅IY!4L/}G, OUln z.-?j)2 "d%GL;xAK ӷbv>-%Ȼd:$.W q/ڡ o#.0#ei%\thZBN!8$„,f"pK[ĂTTj^TbDH,]_špޫʔO,''1{e/#з2TB=t)N1pwHVϋpHHccS{6F}#D lŨy￞~4g]oz\~_ym_rld؟r7#`e`o,o~ \H-?oX[ "O߯sOkpy1n8nyyCO:5?C/YɮmᬺdH{_]2)ϐ]_iРhn }کg zfs͖ji1"ĝ%hVQ9ڼU1tNΤj>%#x.#"$LUάc ~B =fIcsW '| z_Tz-ֽEy`MzֵfE #x>x<< qJtZH)\,eDH{͹˖^fҌ#+D*9*\D`)#=qSPHo3^ *,]8`g)F44:1F|A8) hS:(2-`q[ b ]"N( ڗsX rg rɁZS$qyh}镻?}J} .#4\/BV/bW"墌L=趤]qC~0AH,Aoıד38pR7DkqF{/I+Kqa dAn9O`G!ڲo][T-j")iX)N_[R#N$9J8~vj_Ţm,wg״j8Xv3T"Rʈ#CZJ.ad6|Ob87(ʼ Us(⠖%Ҿ8:sHGq fR FŅOڈ`(Yc9dpQdƗ8QVnrbԆL5'qbQT.Hx v"9":18x R2jޘ. -i '/3;DØC݌6"w:36ۜc YJo=": h^m/ 0z)V2bS hˇcf %m9Cqf`}#8DU6wol#3._K7Za08O泛vӛށ)䲈peL7̓ؕVKmސ4 ac5ThP8% K^ bF\c*t5Ê4uO|TqT`~a$;whԘCIFV s)fy[a)DViQ(L7nMDJ,VDZl _{F慰s3 (_Y6C>i!j6 PB1nNyqGGrt\.NKkK*Lӷ-BԴ #i-v(#x" !T4Kbkeu:JeY0/l_ƽi2q _i2Y():J^˕z+[ O y=5^*2mWFCb/ 8' PNiZGD|t(5W001<D%5ķk!_]24~6Đ]_"}q[nکg^K7_(ޚT8`rW 7 LѬ-Dj/og]& "gTݝ{+.K{ n}%x<\5`}ׅtJJ5t"lEښI'(Vc3q[df:SMK=MZzqr@0x[6@Pپ\~~F$CA;a䦹owowed \qA~V ^jtU45B󜊲wŷ# F9߈T_XC4j, hN֋AjbNTt*AcA_TuqOC{Sa``IAmL[N&m5bƪqLXIה$KiA0> 8f,76_o+Մ`8%L&(k\Dm~yVBqĘ"4k*ql !Z\el{-MwvutaW6z? K$6Ŧc0mԻ~HV,v Ƈn!SSL K4>is+J8݉@(Fx \.6L}F7 )ƻ"W-.PaЭ6UTjU0ێ4+5cfXi޶EةDB/e|+=a4BiurxfA;NxgVvd Ǫdcߩ4|RZ)d2tsPGvq%nk_C%\p%fgh0rZU$ƪm ވ M 3mc{e8Cھcb0`EXc({ϱL c^8w7 ]dJ^G#Z(|P%&}Б)>o}zS:Za 7JVsO*YKf۾2QD3YL%T U̳E~P"NХ~HU三/>VҦU^ӂRKR-^t!W=wO_ݸ7_|:3ٻ (.;p@LV/)M/\%!V}ф%Z3zZ ({m:N:i2p)%`Q'-J%paeᘴ|BR\<)u }}%ic\vٜ>}Ϯz6ÏkTgx~XU*h o|>I)2RZ<-(ϟ>eui'R δA4 \J̟Pǧ07Y oZ!j-la!k4O4Q)&&l<Lس?c# gmD+T:\r=$ۣ=&n K5(鞘%)VցjmuMI(E`/ _IixA 4n7Sؔe8͐KR+-&rr1Sr.2?4!OTٻ8nW\zۇ TilUneni'?bɹ{(ǎmK43=3 xp*ǩ#95e6U3ö,:zϤ | "5|wyނ|ɒ9u!. MƤJn}[(䴆RN?XxRr is'8K%}ԈRR\2jRB9uClV`1 J2([izd߫`i8#\8AE%VfpIыǚ}EXԂ %\=qY(a*ƥjT@u'Xx qm`WO}oy:5Knt~Y*LA{{Mh ~aIo<;bclEY и4mDZ"'-qSs <0;boAjύp̄?Vظngj/Z(Ӈ>:e .D0o,o! Un'K4}k]ЁqĊYi`bCp&3&bOV6lW9Cؼ#~qExȗ-\șy-#B̋V‘BIȾs"9uA[JNS[\h5tswA\X/~I?ݽEm,$W ) CZIPˡhr0k^ L5K/>8%̡Z 10uO߻54S7-+CMÓQ'!9pZ~UZ6AЃ)f2\zy.1:^]C8!hC+W^l1Ifv;tR!=q.V=t$R4Q]&0U5l.yE @l_˚f:<.Q i{vRW]J'}@YGWXDUt̽|\]tsVߒH>%u)}|$:TZm61(1"L#^(p |r'&6\Z':N ?w6>*}6SwO-msXb=_l<p}0G{~]Ȟvv) Rb$4.OG5,u7Ú:IM lsg=&ZCV +D# khU/_mٳ5{V\9p`0w={f.././V+>%N^ٶ5mE6TٜЍ'"ʼO*w$Uy)xWL6 Mir1DFS` M38^l IxSo7rȢ?lt%*]2J[՜o]A`ug7(H&f`ףZpƉӦ 6w__n>;ID"'`| dD$ f ހo,'y;gэ}"tۨ_K06jw UתCSܨFn7 ɹ |rTo_?9Pzr&i{n3swF[|YZ11![ |qGsr9>^ߟ_88tv1!~srzǀdz:-g?g=?kvW&˗8~s{7^ POLG(3/,y3Aѧ{:M<|ø۔pKHף*0Em{vjl0auؖ:[MwFMj5xȄo¼3uakQ@_YWnf+/J2w|_T5YoiO韭 INN;+7M JèbI 8[CH1ۄº?}C-gR@I|7If:ojJ&[tꢋt$3sO2}fm' M3NlcD|xNQ nIWBZ#tML,x2õ,eWm@4RHyH[Tg{zAn ؝2uZ9;YR-ܳ՜(PnB[ʳW-5u2u|0f4= 2OwR^߮+-c]ve_0E腖Rk5MƵ&ɪ.Yub)zJAr/qJ|Qr]w\*_M5?*eo ~^zU0M`l&tǒ7|fA|` !h.C;gQhpR8I1۵%*K8`7ޞ޴&ՒⰺzˋWlRS{u VL%.8'(^AG)z]$Clsȵ+@:+7={`$S*RK Ǟ(WV#:ۥ5|t]lqخߗl mP=8 Vr4ۓG~e;2^O~rvgqwrB\qՇRҷLdEnޚޤSʽ}}ߞ>>/=;=}}ãGzC!=89?8\9y/{=c.M <X^jz`ur>K|'SiΑG?}0wP܏x.o=їOm]W#1}ڵ DoXiThpmm#g SG{42q,2q,8kW)T/1oy~>bKzzx]] *}).gB/{-)efy:ܘ%sꞭmS0:?6t='@.[68Y!]X,.]6JDzٱ1[Bϛ!zFV#@L:YJGqe1sI~zA0m/,Rӄߓ~ƭ[1ii &V[| ԅ+Q9>R"^Q) ,-5jOf@p:?[iCn9|66{= RԡR&7pB>[Δu~,(O>;mK}ۗ/nBv*q+y@w]J`xt#sS@ syoEeG ԍ2)hr۫(Eb&¬S`GE^v$co]03*Gn;_I򎏖vڒwά rBa˖oi2kJёf]jc;(D &@u&Ψ6KηSSYsV5H 2rj{nNxV~ 񶅂I[lE>CS$l,swKՃOݫ_{ӣ6W|0*zRIZyt5RFʣ\a;HăT<^#o}Y‰vosca WW}G6>}dCM{t'ZMj\jkyRׅ{KhԦK$B=2Z. jpp,y֨[^ƤCk7JN a[1% 0RYm":> fHӫA՜ǽ=#J1PsjIbIcI d n較blnMb r8[S c^zjLQeu<==UHE\3.}cf W6Ф]Tұ DyRLunQZ C_ʶW-\@;m0 XqzåTn3go&Jœ+ whK{m^52<`m9qm޵-q#_LI@&n>l86pe77 ,R$+u"VPu!H U>;=ao{G鈈Kts[f{a=ĊѤHb$aH^%G}OcVX1qu+ܷwI=I[ҏryj'/njbne`hk/5sSA2>+%QI=f&m0W.rs{xwDnDo< QF<v7o :95 UH 4G%ĔRŸ%ovdk3M\aiKe{Z˭$̡511k2BeRDG%+doX3kEdW'f 5b Q2 % 1[1}W[FrٌtRhd'#+}K{_M@3 ǻ:f:aJSV'JZ:eG ltoJ]{%ޭjni}h3vwҢiEqvbLJaEch15Kuj8(MnB;Mk >@:\lpja*QS{zl|d8hzL_g#Ȝg[{ٖ&hDd6򨝸`yK-ϒ`'\^ $6K7 e/^GM*F-$*:"=f*]pԗPR/)wPJKdX6S6~6j.9LW?y]@۸m5 nSvl%OwU3QF]X쯮J24Ԩg&FbZ&O\︾f[Ƶ}&8.}콋iY/df훈߻ˌ5/emUÜhPͭv2ꓤoܴa ~bIc);Nȶ~Dގ0iq6}Q,6jDc6NA' /αGp-罗t>Ƌ8s\xw;W+pFY5!L4d (dQQqE5iw0T (1W i̅6\$slL-qC|Yp qZ$HYb#Gٶ6N%kO-ٯrsXm +n4LlfFm:!^RUo̱Ɖ퍈=4qB- %s3! BO`l?b 6@㭫ٔSXZA Qz)j "`u5Ƒ%DGi}#B1DJxLdGqINPܥ :Fp5a$=@YHɁ}&wlb&$j?11c7OW~߾~ H1˶ۨ4߲ƗN]XVW h+LE1ܤsd #jnJK{qQ|wwʨjJuhwގy>huڈOޝwy&yQ#|Nڼ[{T }PSݽ@aZt;9.o~H p]w"ݜT4#A.A#H-հ[@rU(Py.]\uF%JIqA 8>{8'q.o^H1axsDd? wM ͢DG\^g|!;ˈ^v c}w &91 {oDH`AK^svٻ۲6`cT;Rg}D>޽$g`0i`{<?Xb=jx7(PvT)A1:rbIr~_h.k7QFHtD_ϰ)E(5Jq3.7#'հwF܍y6'ib|ԋSO{IX'Noey6)v _ی\IQu[K|G5# ɐ, wc Iγ Hg\n$ux*K3qq7U,Gԉq#Pn<),\ jokխ$g HEV〲ro{`RJ[Oa1Qg+ܵoh6rY=Oi_47J)_=]}(YsOGH +X 6î\鉹r shg8 h7CyNqMhXAHϩw_웅%N*w(el}޾y᷒>#m/bo׃;MzF}ÂkxG.{8[dְouNyXC_;J`wg+]}cOk%V~+[Ơ/6c!h$:F1輶Y'^ hgAV2s`Rp3y#8^xv+o6}ΖU^8q 2xX!gֲ$o2m [¸q_mi_ ͕ őxO_֖Ó+^s1l^-wtw;!~$E_!8ϟt߹/f}RT}g?Sl7pL=O{OSV)ɣŞ1՘mz.̵B`u-p-B$#3WTVҨmv"Y <ç?|ߔ/LhW(c=rOm +LN!8"L5T:Xt[hr K5jZB_ܨR5Rլ/ePP{ך ՁV',9AruQrf&1H"Mgp'W?kt>EC#Dv|F#ze)Jp n̳/cvCFzj߽$gimW`h Sw՛jFEHA`(gPwو9pK{"&$h3͠ 2r]i6:=3^nY_pIxtUl|=#&LB:jqh)M S:E<<7mvnK/F +M֮YlAi I zE'TCx6q*,ڋsw3KR{i#[8%dI_1CJΘCrɕ "h OQQ1j/uc/ g͌xf#q)}Rh'5@3 GB'=X|;/buhVĖ]U DdX* FSĖ%\яvzA-qW}jF6E5`j՗~Zk5l0K l6cCF:5'p l8dFf֣K}l"2 ewc)S8[ED%3Jf) fv\Uk瓄ZcdwѠr>S6\os YDFǴ1(E/!$m.SIJ!²;FGΗh{wVꏟ"sT4k2g2y6]cTcäC"MH)+*ur&T:qpʧ?eNI{78Hhkx O*/Z栋$:_'ǻ&qfn\x$(9F3~&oBq:;>5lsrxƻǢ>}vSj܀ğnٱ/f, 1aU)FZ(JGt@r!r#bXU$\*@^շzbDfSԥ }bLWnRZ.ͰxRl-,c ĸ =X==>0:~MB8D4ػh+bP}7rMC@_o?ͪBJӗ_7~<bs┘ F`2ai/œm 0ںFo&:b8@ɖ% >ʴĘ#~qUS51=k7 [=kx|nf֓y!/AR'p;dBDdjQ=y>o(Ao}QaqN_ Hfxl% ËW?U_p/9iLB 6THX4ҔL 6L 07hC.5s%l ͧ+b ̏h얈@YY*1*(ۯ&93q:lقk R.0"Ku-VCoq%BJ aq[^|3)Fj^"|붇mmbۼ&åTu#ə\0-쭨q#vsٴlE^ NWB[֊# 1.`thK@jL]zh scj8ox{sY1Ux!;CpO70c2sqk\>omN\M FYA3%QƘ {R$0"F-Ɯ{3J9_Խ-/(B2o746Jq UΓ r2CRh m.vz42R`\kmI:bZXaH7K7]F_kaԼW׿}pysrn:G[tZlt OdoTѻ~oC$>ܘ&i"eDg&Pcq7<0qa+9$՘*qn9L#D:wcF( Ϙ S\cyEi5O޴9/St&vθ.oW׍-1;ms٧8,X^<~Kլu4}G z^7v4s_ջL@<ч3^Ut48J;V13b9rkM " -%i.Z걘DQ8,z"O 98W1jJ@)PaLZĀuI T9W#gsOu RZ_].f& } 5{[.OuOߘH_Gkpߥ/ۅC/~ܿT>oh?]_/%oAeG^F2!l'sg)SgLA=ȅ,HdmRFmFm?lCNSs4W !FE5^FB/OS؞m#cc5xm=ef(v)`t k58\*;>' NWtjE0iqTJq0WlL%;]/h@@4 Ѩv{ş(ֵ,_#ٙeOωgwbm60';{=;!';`nkbҶOt?oN_,>j@1Y'v̋ B:b+v㣛[|3GnxKo+Ƞ 9V+b|<DTŒ*} $)L헒a8 ђG**gv}114pעb"5SLN1me '<(z힊 A|KZż!}6| $bMˈQhGqF0\sKKb9Wge._ (kRa?-^TYo"Sv"ޤhX|ȤH~Y\ogB5GO)f( r`33?!w3^򢠾Yo Jp1/ |0W3K Ir7%FUZ=m-{It\An1/f Yg7<~HrfӉ)mۑ 6]$'υƱڼrջ-TȾm4: qҾr9f]x"u6n?}:1nRGf-vAae͐cѼx\݆*9/_C$>=&q;A#ak7t=!`ѻRLgifCg|5G_Sly!$(85+aM<9?:ueD~U9&z$3:OppX>Շ 7U%%uYIe]TR} u-|FoңgBlfe*I;/3iYrG+=o#[FR~Mt8TiG+=_VhdΖC;j.jY;j2 D!nv3H+-#v8y@tB}[fO_./ܸ: \lsw'a:Z$JAHA*jK]=f_|أ.;bR)~<ȏ)6hvў!eV0 ȨBA(wLѧ>f8ʻwJvr2ig^@썖n|IΙmGckᡐl0F}˦B KYx{s-kQllۄ p`u\'W"悘4 Ж,=o;yf'ċdQsPNA\[2:vn$%2&Q*7-Dվ E)) !Pm eXjORȍ|ƽx_qϣ>ɦ)l_;68BE%e6X+<-ZuIEEw2 juQ7) M7o~~'o  Y2UInTԏ VNرHrV#>0C#%LIQ5";^L%Q!zEjQZzȆi-1PA^/=[dp_xkoyO͐b#LX !±(V3Tv2aA)Տ\zhIL:=t8Ubj{̅Lk:|ٻ6#r^FC?kes@kz!sESZ#W=CI#rHpHlGwUu]tРf1/͝e0Q LKB$f6GgFxtb=VRؽg Lb2FY,UȍB42t##-(R9v#0y@ѹۉvyˋWR.ڠ[6L!N-ҥ1c-vinXȽȴ-v,82g Ùju!vO[1o\SBu+lW~z3i,uk & &}Qӊ(]^+ϕ ;4%AW]D d[ qt"4ěP]젫D XI17L!F&1!;XY>>-c_s\.e@F>wV2`x *Œ){ѧG:7yJH)+Tc]sek.s͉i6e^8J>+B׼Bʪ)Rp"(YLI5&xsi҈3StoO{H=FI9QQ{)W9tG蘬Fc0jDK.RY?xl cPvvq>BU78P SӪPT70݋v-ƔT{/] lTGf,0J ZxlJK.AUO NvVqzg( @5/\0hbŜt.\r_Rri-A+,=Rd1<^2)*eͲJbd1$VLÁ[ sXd1"ju/t*4,0ݞ@;ASX`X7R#!Vpv-Q,bTI&kuVu;Rhǰf8p c1 ̩9MU\ SoTB# k1f:'UAKncD%% _{M7[on2s֏0$? Ec2Ru ˢX +gbF.vWgͨG2% %3E= 'aBV" cƍ$NDZ1,ۘhQ. b|wwec Oe)r̨.6L\D6 "W>n}UoNh%[=5lA;N*J,WD[,ײW??gٽ.[ş:(Q6̬c|34Hn8!.?'( B.{}nP!v7Ԛ%;EcQSrrfť5A*$6=el˴O{>(@2S:D+(d<u 2.'$Dv$+ڗlQ`*>'@Ĺ *5PkXo'q> `|$N:ǰVJv'J$5dԄv7^Z8]+wYM|t]mNfњ<8cګ Dhb+ $b} D 9F~U6X$ ffty ^@V;[lyJ>$!:4GJi_!pgo=`B/-e`_#-yKapФZE ,-w:)Tw(BUN#R΁wcΓL-t" W*;T֘jg#4别dI>C(c̀*c&u7S^MxNɵן14U':W|XmyCKy3G䰚CH}9\s|B?`v-Z "_uNN>+Jedߜ& .p4SVlٻ9e@j/DQuLK3&EcA %r^PcJjfv;\$+'?/l3X* ~$6F2C-cyt92,ʼ⋢c,aN֯]Ի 2iMԃ8S$#NZHG8?DIW/f|:3<-7-N062D``$MljwoL0sE0N&*\1!myj`983Y2CΜ.`Cl;xdzlآNm`&0ϗo|w}`-yXu7naCb3%bQ #C$IZSBd> MvJP$/[h~0Oegq?_N9H,*R Z%aT"Iw&g2J?G F*v.~zCzpy9hҧ vJ%O(e=sh}} -ې1rsyMsk7K"fMްuTU$d#4E-<$=&٨$o˛{nمRiAV3@HZ)ӛ'\vh-0t6BjChDHD NSਠ"eMMRT&<87Fڗ.Io%a|};|0icZ3E6ھC< 99}RnoVGˇfWtj Y}̍2ug^7fb|'?tid&Fs*`v4j~F(1M9t%3->K,[Μu|3s))W9r\ۅ׵vh(=.סOOWy6~Ď ܷ 瀼R$"1lQPrdROJS_ErD S`$ԔS챪@ੋS]t0U\e5-ZZ?SfnnH N8~g )1L2 NR} 3cSJ75ŚTO]0 &|; ,ȬoDoI +&%tJGsbb@9<<P%:xЌu! dmT"vɶ-pyՋ2!~ɶMm~27ِIdAeBb<r@CQ{YCuKrO(6fakӵ@R$0grTS"Eddg,95nL.ۤ?$. %92d<K%s2Ƶ7^+ f^MD2pwcg(4,Ka]u57Eot\}05)^{&EG$"-ur04T}QG+)bRf]Ħ #b:yw31ۼh Ϯ ff6S[ApL: D)E!?Fwz./R1+` Ɠl|w;x!tVT3~.kVh1MҚZRE+PmLRb^HeHa\)DFj1o͟gr?k}N `mt'r}.9KlS?j3m Ƽ-aQ\)WXv>D{_pࢎ)#(L9*Z e}Z/zQmt:ARZ* M1V6Ӎ'D1lzZ_`eޣڜٝ›r oA=$g4Z m?G4p-\9 \dל4>ďv{Pm;U^t~}_v޳Ojpsw''']&6<Υ8[tJrʺlOq`R AS9:x΅ +07heP |놃aҚI=z+j f9 %w|3+mڙ#?ѷ}zgq2r5o7/"_ ]ijBE<+tQ=+4{3&Sg2R9<`Վ'=ў 3D84@ټ/XpsVzbhwvcE-,*۬״xgCzxia4 \o.͂(QF*vTt:N-VQ`]{}M ,x}ĕ3(Z33 L-t_F8_31ɉ|j] t7?!WBôx݀[ CX|o"ˋ./ >FBvgL x1Z[i  4Hx9,/>|sV0쎥PqvGM;" U:JUJQxlLI)l&H 2kkV`iKRm"Ĭ̛H3Oϩqb+rZ@*tyBQǒԈvr:S) QH8(3sŌ;m=L(KAx;Ke\Rps]{ [Z̭F(\D 2"A7"pT0Ơ<MH r YZ{Ne%֭"Fr}4\XS0'I$)b=A`F{& "F ztKs 3Wu?3k4;> jFՎ v0%ٞt<\UB=Ax. PI dEnJgr1Jm e@DGOJmV'FC:6ᠪIFRhMY/MawEn'?{ȍL!i`,ea&0XEҭ%9ܗKVIVfUl+tBd8b梽sHpvԆ7^_H1EjIfx[рYLxZPJ>"iT]4s &=N,-(2FF{S~i#I#!mHÌΖ ~\lA=_Q/ëL"B?n6'iS]0QA5۩VO""ϗ@0gH $DJDV?YUALl%S`qͷCK&H~e6 =8eAԋcHA)C T,<4DjZIrO((%_#!O|;ÄS%$9{3#Z?(Hx_?`9L('Mwd[out?1U-;$]ow룓q0#*Rz~J3o8 (H=kFAC^򿱕?q2Ii|2wM8'sT5M` ԰!B? <]d4-MJ6ЈZa0Eqq7ۭˆk ;|᯹ ,?Y ^a1p~6I+P ъw 9? = o#;Bў~Ӆr.+>xūu0E͑.eT"hd. ?nEh-*\!:8@^y]w'0_2:Ď aJ80*"Q0 Μ5q$2:DꃒM>5URvh%swP̼HH#35`BT|q\uo܈(=O3"Ms!Cرk/Bȴ/XR qXѓI ԓ^=uГ^=u]OZ?w~cT\#JIrC|o"'gl@d}^>(N}xkfԋM["uS̪rAa0c:,E {.*qYCۻi= bz,~% 8>Sc85*xjAKKgg e?ckܡʅ=c֫ݙ Iv,8H ("d}GN<ρC!Lv4h_G^`2 Řdھ t- &!`oLrgߋi)}M|ha]Sav5`tID_I蠢\+wm̥c<%|R]H{*p+KCo]Qr2;̚e|=\ VF FHU^Ѷ`ؑrFVcĪ_uwܢhHO2ݽ\11tAB[pLJQas }^TX)zh[t"PLl'JbsLL0u։S:\4R\h$"~I=iAB*!`jBqmȔ`y?Jhm$O]8c4LAo=Y=09*=<ªGYMq*3<(=>?1o3%^%E=xVhðaݮ0jDBbb@9v<Ϲ4f.Tp#8Z` ߒ3P=gj$<%s@V,8;U l]h6c BWXVʂB$¯ rHŝF.+ h&,갆X3 ,#-)p ED`Q %^o2W0Բ{¬s|"jc~è/}%J&%>aHd[7o~{D÷? G֏st!"⯡?lX'oo㛲!?]꿹ճ.LP[n2"R~}1zW NM}tHrX$GKPu$W~hDA鸍}I#%?}/G9Ҁ8k뤀D3(v 1ٸ&#%b숶j ` P,RpD<yَq 3ۉ6L(7%Ev3lU=ľ~b"$$h=67 [Ns~WayQfp1CLm-cZJ:@{%">1Tm v@ԅO`){uU~~Jpų>y{~Oj㑃:@1nx)3:gQ-ekyA2((E`څw\p$WQ*E-E!}NaT:LdzA+< !Y7Di6oh3 Ҳk_.9gX[Dhmz3p/t\T7B`b;E`pYB r-aXB^a 9rņibXzk$wQl_LLZL Kf I䧨JQ%*J8,3t)Y=A\LF=;b;r]j[ɝj^lTO0d$WZAN&Зt٨)p#aAgBx8Tn3]K$0N@_HP̋6cLZ3nAu;^N~+XQGN5[0alr yS*NkgDqqχu̾UpzF=GBqDkfǰkcif^dm ͼP$U[B~n ruظ}]߸rmZQ(!FSFBp]RT/HNKAܱ>nX"(x"[&(*";2RExn%D[Y"(@cV8R?7bFv1b#Y+ mx p3%BC:d!XHB$PKr" 1c`)$*$6Ðt 7gL cصP@څ.|h'avsS;xԛ~0~_CzTBT\$1:htP[JE54/zx3<߇^=;[TĩnnGm\\^}<|//> ǫovۋ ?o@3},mGynvOBAT;UO//goѿar7 @J?鰘erˬR2Dr1E(8FH+%Jjm^H)QOFp]Ӌ08 Of*(UCrO/1.hy(*M(PIBA3>YBAzM CL)-mƳ!ܶ)N3-6.еy\l ̬xnCO^U&nqZ`Tcy6xŵa^q >kqxItmf P OUk2 ]6^]gzr pGahwݿ:jHǎJq^CqصYC8m 5Ԉ5„DbGٱ06k!gv4;Zby Eaf Ѵ5DP#> k v L9Hy&6s&ɳ_Ŵv-v.wg9 r'W1sMm3IiK(zzNb `yr%(ϴ.sF2[5޷7Ў,%z- `&ǷmfP9t:;%aJɒjd6y>jEfe<=16[lrXg;*dQovsN.RXsMp6~@ ܖy:^Yб0vDD#u..$ p;)Ce¨r1%/)fXvF;1U i{'Fx}1.*ֳC* ϱ_Uvm*d0U! Yըr /2*Fj|8dDןT q_?JT%v<[voOF~`_`R·v6.LofWʲBu }ȻAb|n߲?Nz3^0x~y%|ǟo?ʨԎf|7 ?޳op\s jo.I6ߏo^9};o6H|ŷɝ#+?dg+sF$XO/2)_AW=C!ipdb8T2t401}ynqs\og=K =b fWzo}L7ߔj~~{}tʫ?^ť砜'wQ}z. G~`vVl]0-'oҡ/C X&Iӿ4.L _xa23<=sp˱=>$:ըFO?w6Qu_яK3+e]>ͫΦ\TxSIW3)o\ɚӯu`nn'ɓw?ip ϯ^{ATJ"UbYc;7)Xz5~>>jz.UfbWHԔF]b׏)Jc_+G6O՘ծɲBysBYYro\,7`$-@ !q:+{YF!|ۥ=a#̅`#ap9-/EcU}?Eۓ> qi6n8QQ2]>}Z-XeK7*7j 77/bLIO&ٛ9_؅L/t:$Zp"VcsZ~hGlg%wfGke mw(7\675& ?)rЮ/R(ҞLLUsxji~5<^w7>zmI6v<Nǣxc/?>#1H )X`"Rţ%ˠ[c)9 1GX6gĘ]5#=zG/Hˣy"^]^d] 9]T#I4fhRXp@(xIˏ8K9%٩DfKPZ; rSϋ'=ɣ'y@#;=ɣ'y$/)i1{\Irłɀ /'(hjBJ_]'<5_2C^IbԵ.m2<* dC=.<c%(Uy) e+3H`)EJ` -<,,P,j̉ րVF ')ϞrwGK(;Ũ+-謉Ҙ>1_\8\򀸰f/=/b#8!@چ _^9GfY jYw֖Zaǒ10 [bdFb'9kfX"CK6E9|p]ի UHS.V7$!r)ClS>@,mzc\\ ]q%S8-Hnvs6B|qanDv35-DsL+3Jۖ{l -7mnx˰p`iB*U{Ҝ҄v( Xh'%VP%FK}K0-5&#NZ y#5Xs} C<BDgW  ٕ|2tp"]cp-,?J$\ q zZxYXcζ%䵠vgʴ,i=H%zg~Jme jMO[wN<-Q䲀$..cҵۚ[ 2^˗;@9hYB[fHqJH=/V䓫5gVJuCgwkfoh6txvϾ/C^.WX#_[{TgV ]߱N8LKp:;5SGayU5hM^L_1˥q˰YywVz\DT>ġڭMG6mK2Lu\D;ɔ1,k7AhdNwnDHNuhvABr\FZo͊WSmxgv5@3ZЫZު/~;"UtM%IU.E0ɠ]MNƗieq~I.03~1~ Ȓw7+$/_ԁ]] &EZE%դϺh6uBaձFdk׉)) b[jʷJF|:x;^V6,W0@ϋf:y'NΊ> jGӌ\7}֝5l#@aToE5 O ; BLoPtL󑄂 )BFf)Pdgm'8ֻq{lX<{i!{F<}8B4Qʪ{'U(k]m~TCiuka?Z(@SLVҒ⫑oJ(]Q~,bkP-V0oR+@,Ug8KR;NI)V"0pEM7I${Sp\p#GNZ@Z{i7^ C@#Vs) 0."yk"SYJ(3nWӓyCL/o^fʨ`I [;3ݔۧsMVAΎ-@UyQ{E{b2ic QJtЁRBZB[?*XBtM] oRw)HHm)_*I߮$♔(EjXAql_MΪJbm=v8ިGZ9̆C$ɣTMdZp"a\S="5OxpZ7Fq )X rcX㨽rDI0 وb)Gj-Ac/%BK4iZaHjCxEע`2D ljjBYB`Vi*\pIL!A51.q뫁aO+ڻmSrzNW4bѪ h*ϣ #1H )X`Th0ur"3TnCAIFf6^+=B0TJbX).dsR1\p! ^xyBst8K)ւqo8Z_c {xt6sTݞ?wҷEhPEmNׁ: >zm: X?~{[(=MܜFp/2Ĩ>2ٚ*BiXM`h ӫnMC0=1?8Y, a> N=G#В@J0Wےgڃ+ɨ)%UjW2]>)5 '4O.EfzVIaz'B(VAb`KtDj+zAL 6xDڠ^_>o>߁J![ (Ё׹j ]]PAp3K;yecAa=ϡHrמ:DT^S< #O'vF{ Lf>9W&2f_OqDIq# ڐiǒT˃83EDkl76)F)2X3c#)biщ%6hNkO©]V]҂TWQ-͙=\ [,9:s> [gn F`RD9C2N%211(IiDF7:&,X)&"!Lк4Z$nm%P 8|')_J(mn$߀U<ކj+lx/F#KL}u$br4gkmT9=rdzQ{!qS3:;4 9 +Βy/lss, o|п];|[3f+"b}a"!f֐)v6ɽ/O1dE=.0)‚4DiLB P08it?f`˰lv>w2a*Y3SHg60zqWxdIݤA71s&&CNGQ[ϔP`E01nr91e>OQv~i=l噒x)UBbSϒNeb^s/;sgJ'W9?LX{xgʞT} $8b:ϓ/"EgTs]+z&b 0x=5]Iϔ[23['4 恛H-2sladszgc>ٙ*#y! =gS>xsV*ox`}w}7߇7^~J;ٌbԛ|}-c ,[fF W*y=;vi*F߅Ku2;EѺΖ%PK [dpl/chQ;{2;0vf:S9BrBIt:3$ ֫ɽC"Y&}pq\MG qf/Q0AR{L2 [{so^Uso8W&2f_O ν‰k?Zp)#…9%߂I~ &SrFdjDFd}Hn)X%`['&?n7)fp|krq 9~u˵9#GdOQ=^ލKubd V 2 uPX4f1`L:x7*}F؟r}LlAM.ե#3RuM'F[668ZO^|n]i6x<ط4zk$K:BlBxq0K]|Y^&?#[&l軗AIfZ4JXJsfVQ# N8B8F70epl}iEn~7uLc13@_'nt0۵ g5h"Bag~?(9D8j%WgrN T Δ<\;\vYࣛ j - 2'53^j_!||u.e3B4!q"#pAd AE`Ң8SŤM.Jr('\c7o2.nї$+t.  b0WkL啝C_X?Ӹj͐bSkNf߿ۍ^lBA@4gU,0dvK4{;jO=fB-[Л$YEm͝*vLmi \O APx*Uk+/$夨Tb3@h B$ z",M7H"(M`+Հ?1f2*3|j*&l3ĺwSQckbM )A:1F0M-'*2pKp_}XW$ђYys9{9We=f. 0yyJC7o7)ٴ}y:'-X<>Dj~U.㼞NVt>|}pL 5,\ߺi bw] Ϋ[,=38owDg2 z{+y~Ya!"BD- fSFAluލWfn?{ꛍO{lOwW4 ;x1\L8!gO1dwVB?wsjKU OXmBɥG*V\Ni*s Lg+) ?>hYwG1?Yy>UC+|ͺ|bwE jŧ^5) [n;5[m̕_;b<VŘi6f\^;?"1չVUӻ% YtV*ƧqJ4ꂑWXQc44h 0d"20RTTbĭbt D[&RØaHѳH~НQY&e~v&J DIPBbֲXiP"Et*Jl"iRcFbH1|M#ЋRsxb(*ګB\^"8VX"J00QX$3BX"cP*-t"S%RP+q` 5HjYX5G0sDzًr ]MF>؜-Oh1O'紀t0[}$8r>'jSsm~m>X en:AE6r]n.%ZcUSE:Vƍ<μN1T^ʵ'ݖ%lKWF:%g#y֭)>Ӵ*ӢP"ߵukQ_ֆUNյn=z֍#R֭)>Ӵ*QGumw׺!qSCZ. K=?USeW\`&z+ ]X>*rt6&Zjt`SD9ŧɲoM|r/h<\ /_x/Ƙ;ʴ5r-`=مV#I4f1b8B d0$ X+Y$QM1II*8JQX΍!HIBFFk+d"n3YF˩ʳ|MQ8Z Ri_ jjC0ԆK8u+GުOD,*xIxm 35^אVS4\sb>.0JJcf%P5?V^\_?[1).mm*Y,* w6VAbWUѴԿ5`{z؁]m"a$d0: 4%Q8 ET R YJb)e"v 'Z&ai8ɄiV*ꒅ5ݕ5Е@HH/R%_9n֩La4JrLӿ)bCs[zW Z.?Z*A+)`lp0kzK(KؔW0>`hcke@%P!C +Qp{սX.F[?iK3N {/Kw dM661c k)4Sk^-[6Ӝ\E>`[G95tέ7K#OJ?{Ƒ>:'qIEاl߇J5O e}J{z xE\N9=O@b5+pX*w.\H_Bb\(̘R4쒔@d? =YfgjҚӧ_Dc%CH 7k.wZXD51=3+Ź̈>X6G~5"vs+Nm&v"q^9$ &8Pk jb,hPJq"j3[} UcRӔ~ 5A0& !)%2Rh0ыHeCgk,PVrS ~jJ#mebP!'zZ v3è3ր_~]L@ގpc]Xz^_zi|9A IJBfAg:mV*h%߆n˷??kۢ5"6"J$MC)/)bJ rZ6˒CH4rYrZvo[K2^%/ڂ4dJUOmF4E6=J"K\6^h,'NoI7`mYL! u8 [XRYӖj%pQ#FY"xIZ&Xa2W(g[dJA-]Դ]: +&ꠊ+5)w?p.2Ky4H)ʝdk71"#M8p8hCldD,JXa2kZɐC4j\\6QY(eBy@ۑa0f;a5GWoiniPbk&8Ǭ*)VR"0È'cYyEx!ɕ`{~rc%34PAd3{l4*ztwDއtMQFiE"lq-~~JO=|Z<<2w~0K#ݪhGn&ݛAN-orHq0}k*[7ũuU/WS4U){<t*ڳ--D'`;b2C"JČa`Gb$mnKF@}Gbhwax1pGҊKk+Jfb~cGK Zuvtȼal28;JzъZlRU 9%ga ܍/hGa0[. u)seC)RBèdXafOӗ4j:hMujon 7aӲ/`uT@ "J-V7LFLvE([]as:M[.f)>Y%gMNه:.$bBv`x!"E0? ˒\1B|LN>vL7!ԯs߂0ttn k)I1~p;HMzvO~`n:37PÈaLӬVLRtwaf^+̓NZkFw?~gu~e߆_}k>_}=2/?&a§YZOqI?WSͫo޾9Llw|t tVk: OkO|~L݀rAǝ,~Vn kD?pב !4::+<*3,*͸i !lUӽd27h$=ٺ5,ja_ ɚp_T^E?`sX73>ADtj& j,?=qA1e ߏ&jnGE] ´|Mp}M*\/>1N~Y 9:mNa;߲^=r6m9|[<25uYN"90 8etWEp.2V(Df}L?RE:(JZXs(%:N"BZB&׃@Rv4$O:[ħhJ͆1G~@z!]9hcRp۱OP%4C3 &!&US:qD=-a&CH vaH;clXŌA _"fǛQCR=˞|%)G5܊˳u5I%QG5C~y>  t=O-SXjbxmd [X \t蠔`fãUqQqբǹY)͹W hWE[ӑf(F`# :2ׁSc.VV-ReGrI & K~jdX4ML_7ͳGynrHP첋lӶ),HbѶAQl\oix|c]RUJ/]U'}eY.bze]9J)cԵk\*?x2JCf(T{?3y5cƧ z7 #ߍs,pV^ G뼽u0W 2?;2U3cUɘV3VxQҞ]2 cü^ba:JW >%)ylTpVt \s.{ q#|ȣ@W O!XcH5ADD72Գnޖb쒾mS|Iͷ$.7(&# Pe҉W@{/S0ߋSm+(E{/B ޻{6-syM)+N&G#ZI+O+OmU \HvFDhy"09;&l}O#9r҂ݤȼKcZ FdHegsf_-&og|r8)`~Oe eIFHSC/?I /O~k,G*8p8a%Ȍ'D,JXf7jJapت{O_5":{v[Uc B8X P5`R0 K6d1||~4Hh0uy+[~ Yzo^7)S|enGaZ|9&o<՘>iWʼnU>g\dZcy.=q-/YpH3,:tt 7FMRM떉A%|F-fκZٺ5 !\DddӺ3-Z\ N3XZR [6U[E4K(9ŦuSar1":ϨcZGipnلVnMH3,skM"D>떋A}F^xʪT5dݲ jݚg.Y2uhvw!MI~-JZ֭[SNw&uk@B.S כ 'I^nMbŅFS{`M-^X-wHzJ, (+c{'Q)k$Nv`,J3Pv[ƵD"a|vZP\zX>D">`cDGOuʊ3[@tu6zbr:G{e6WUd%^v$L`2H)B9'uRcd mĈb\uӨ4=I"/kF~6yHRW"> AI  ԃtؓ/z=AyiADY$3QAmYecL tEtO!OYD:lT+2;g bHa(oIF%.d'[1_o<~ɻ6$W|&>嗝 EuU7 )OV$ @諚|Ywe:CO[(ݍn7"C~dQHE[GA(#$h pY((˥buVkLʵR.gS".lMѯ'oڰ @  _XTd\/`oL }u~ [jKNF>&Ӭ7$@.MsGx`1NK%qa w!_aإW<]s0|Uvz1R.hDB~5gpETcnVfxE:%Os2ij&UP)7NG:,}a({ 5)`̀fېzvIq8(~AxISۙѳ)ᠪ=.Gifzhׅ/#71ݝ?ӄ(hKB 0 ͫjC՟8<Ֆܨ|fxߣx>G7em6n>q{w>Q Z4M.a\agIKpһ"mB?6J uyAtz/0Myl 蘛uqҋ.)[̮kQV/͕mP8y{5߃O/bj[jSj\@HSS:kl0eYO`b9YH%I?ܗ͟>I|d{öwGHfK_';tہ2AA,=ܾQx(&QѧC}p-{?D6[Bv?\e9ZVgN1xofUn=n("3D/ ^wX\?)"{A]*|^@]KD$|y9}c2.Er VH=YlgcU/;%e~rĈ8he 8W$DG$B ՛|^i5OtU2 B ]fbblVb~ig=Ef%QE\9seHR6_{k |܍Gl7sTtOn5'@,{ymu)v<uaj@_4!032 A]o$h9LkקC"LIQ/=ɇ:E4q `#gnu8A > Tn"3*F%13;L1)Ig!5K ab3?eD5P4-!4# a'n]#\XW+?>s6@FEgR K޵Ҁ9 7Xz*BSpl989Ǖ-&LMRѓ gM28*],kXcemʵ 6v\@D2g ں%x(<ռ5I\R-.3_NiQݷ2 ǤDm6kO*,VI֫գ !gyti͟L EZ"0:%&$8 c+bZmZw)a;}x9䎧báIqwab-M%x~E8z +"p9Ingus_lv yS@jBHFJeSgSpm@&ZJ2@mW5.tifΛ~.pwto뜦"5Kv |˯@(^mxEᣊjFhdxݵV!$A Gjj 䉤_õ@PDb#`n**F`E VԥC[z: _1'E?y͹V"E*ÍI2<~Kx>ؠ RrK,)=gRr=_VI=d~=@cپ9+SvW|@ H! s-Qr ӻzT)B*-_`H1XWGO^(Fj_nDC7"](&N'~kOCTS2w3tBZa 1s! >h!u.WsĢeZmмӺ+Np K؟4\mH?j4Bݭ>oP;~fB.ZFΗA-j4~FKcDŽG%#7J;mW1 T}]³o +R"³ x[/鸬/ mUSopN6̥|^bׯ*FOkOo:@xҴzL%C$fL&# >k :`gV*?ө4pՋ#Dtsu;3_P)KZ~+'>n=rF5'd:Ka;Yy XO/7ʏ^K._DDHYq6F_~p1UB68B)Zs*a8eXaԁ9 x2X^déa~/|mf=ehvڿ:a'F<=?}@DeI6T 5f+EK3*duym0սo7bװ՚1CXwoJG{>ٿs{WhΤvhǍb [%: ֒ 1Hsfb2R*ac$~bZ^٫ZDA P-M0+Cdr~n1F6WtWޭ]п9nP懤6W:؋2Fe`_.w\B^^=]/3׌箧z/VH")),ř8s:uqv nܻ+͹S uYpfG̩I##Y8IΙ=GsbqR~Iv0gnrd &-g= M5CV}OMITs"deni%HSq '_٪&= Q$`es&O< AI"4hĦ8V^śL>'Ado5MEJ'Ud7]-7=@%{n?D$>Natkև.0hm|7ur:nl?2OHz||'oׇ@p ۼy#!B]đs9dH+l]DDfv?&Oٻ'xUUc1[LQ ^lG.&t&KAibr8sQmV,W8 rX3˜Rd<*}_oe:a`A6/;d+̧k`.l„|@69I YYi:8'\lblR9h`-3AS4ϩK,Jie]&DnL:HlIeaSd^]8Ƈ8pS܀x" R cJ.B+G,GS'8s9 P= +Kr$KR~Dz@ ᠊`Ot8mb@9I XNs&(y¤fIJĔiӖb[fIiieK|GS,, 65V3,QԢD"eT&ى|4{"g7gC^=90b&aI*KrxPZn y4d2t¤xh΄JMRF5`\pKFA0;" v=ʧ#-+FfMonC$|dexS*hp7(gیcrgJdcxRhef./5Q뙻M3$ja4e],@t;1zBIEBEǟzuC#EK3#/>LK݈e5ղcX>1ml6}<}ٵ'/SQxRzmNͣϋ_w w]ưnS,pFu,a1Ϳٻ6ndWX|Z׆WᔥM*{<9`$P$C;[OcsTbA׍F7&\2459x Ui> Ct$%dL1ԍ{8='̪hQ)#_ r9e0Y"u!bWېv!3qޔlDbyr}cXerדD[` 4"+ro؀M=]wr. {3(& KܲM+ﳾw9rj;kNs9`(J9 j*vj'3UfGDzۇm1u\8 =m Dx-FEp"'s$OTVa2_λ]ġ&x{Uq r^ P)B͋ӻ/clTe`<$ylާ ӆ=>m5vvgx9:0=! {<XB?ᐇd_qy;<69*f2+`H;o5dgMf}P{ FI|_{0= 3pI;wZ+hC6waBEkK%| cZ%k ނB>F0 ~ >Ȯ61?$H/a(ќ]9%h/D>&Aqפ9^o $~fq9s??}Q@❫UanJ0"[#xY5WggeToqj&+x2FZ>U n¥vsv-{*-j:.^ VXY8[Zꃢ;Za/r zP[~Ƨq~A +y]i޴Jnӭd<%TꖨP=~,dt"Q4.w'%Z=ݺk$ʰT=@,dt QW !ֵSvd\Te4)[=ݺTLIݳ}m~uhB)Mv>b)dIҬgQ}diK~HJ} 6px6ht 崿7-|)\^8$6 3FAPpFAкv+ &Q~pZ+(8XEzV95Q@}z!Qpb9>ٺQ8ӢAZ}.iII kYLd4Ioy8=KJ_{OQ;Zc wȩ)"Ԥ^QKkwrJqBis|RtÉ?tTΉw\gr:/_~sr0MЫKL/<6M-Mv{(p&X:8㴜~ԵA{Ox KDI j~o`@1CþILy9U9$QJ<];-Ʉ0$ɴ6֒qr%I$ZVقZ-3FhmۆECC\o$e9KP LI%G*3:eDc5a6P׎T`Ŗ\bKdU'-m@ua s7f{n<7jKl NME83Dc v-5VpcSe#\z<#yt`uIZ聾FA=FhF;X s"HR)8X%0hL@ [b{1:eèpb$(.3,Zc29yU`mb3gbz @ ZזޜA=7!o !-DG1(C-` :0`]ޔqb"1Es¨ǯrT${CU[p0M L3"W*M`4Y`JSq<̍RHfHNtݗDBZRuapxx(L*`y:K x튁{c* DYb_C7%zZ{hQ6zꧫn-\/zn _ t>~t6-0Mݩj|rf(e_/b` ? G\鿹z1Q-} c}kJdͰ5HCy:@LrkI5I %eƄ C&g |DK#5:S[- 1AMCtwI{р҄.b~.Ƶq)%eCV:٪J`Ƈݿbx\]?ϪWu}/Ψ@y3x,>d jCxT3@uYZGҡy5u]\{i;s b_.I"?.0yNAP=8."CXS3d*T(j.\Ĉ V"D*#e%,TK,i󉘴R,{#*!HVRpިg}yW*U{n=@U2b5-Ʈk-K"y0[87*Ί-Af)=&`:+6b*~s縸QĎQQz^~/)S)cXn \?"pzb t)Ms٧h|)WiOJnGQV!YMU4J$}(c4%-ՉaxӴuf4Ժա!/\EtʻscۺiA9Q[w!FEhVp)v6ttuXJ63a4`D,Cg!f;Acevc?G(&2hdresIG$5Z7Āp"'yʕO2C5{z/fbEWڪߒ$/Jlbݞ"wz3r_#fZfIwƎb.+:X~r f̽Jxy/Q1YN4#wlTfӄ>OGqJ|buH #1Λm&3X<,h@B雘ZN'૳t{Lȳf<>?Aٯf'$Ut^BFCicND4s&+h4 yS;7}RȄܫA@kan̆ҍi:8jȹ}ut_݊7ϨMۚӅH{hq;;h;M!P ri>{x|)h)tiIah?j}kb!AҼGy{M;hwlbܾ\xros*b +?-[MLI7xԔx2繳80-(~dn춽4̒eY?^>3[bG\qju127e`x> ן{J(IrMY%YRInqJ+cwE%=I_BB7p?*ܽqT@nr8> ,L.x帎 : wn)(mc|r)|O.&Vef˝3Ǡ,}kCHwϝ10?ДlL<),+Ш^7l"9W&jXO:_]T>fUp8ǔ2~\K* puGޕ&@r9ށcp{5',y)X \$®H>Z5$:H"~'ړD2 {D5A$aW>Z'GĻvHZfaHp*o[EUH4Z]x+B>V(įjr b!{UX(|,^ נ#"vHϫ❶w{3SiԵ!v`p)eZEDƫV4\O]:3KYO]ΓzLgd(g{$(j0{t"c[x[a,uߏk_]ټHw-Bmue˕i#Y]&HMbk4$sx<t2Ze5hJ :e1ԝi(jd ض0!7'EGB/U /ȶzx}+J!ֳFAf UF >^( 0xe>ayqov25OVdglpy}|ິ, 2@iY^{s+2MPvV?!V7apR am-8ZO29)CPMI.53Yj N0F;{.P΃"&;y` ():`$ŸE;G\_Sm7kƭ(뗌Ied,*lI T\?mPuX8 ' 2B'$b#vY{0Cqq<)PG㘿 !N=хsaEN&B.0HB1V術5˘yy}t0;+]QmkӺJғA H+N{#1ܟ3n-՜E!L( Ic>LFP7/}a?WJUJ(KE{pIV~Sa_o_ K%sR!n49?wܑnK?Hi IA$UE: }+ zI9΀`(3Ɓ` @?#^)=QH)z?ռ!$/{oVRp@y?䳕nI~m|!*ƃpr &W}T˽UZh&$9 3* 1oq_[GA"ϨypiB哔oEؘP1PQk8h ZNDIFO9 Eu XPFQ^XfJFو1Ok&JSހ*#c$vW?kI*J>=54Xgy?L`~M1\[c> Nj :#Ҁ+ʨUKTyָDA)^wJДFR6@ WhnG75j*T2Ʊ+qlx@EB@I8^0lO'IU >=%8L ޝMTgګ+hXKG{Siɇ 8~R^D/Eyp:Y2F$7}4 ]ߊ4ZI _2 gаWUW)xeB*AMg?=pcG<K)<>Zg=LZz㨆@IE23ddA wz :)55''ͪˆ8@KT] yDm,^BB^Ս rHJ&jq SQgR=O ]>G8 )6(ъ1#q x*_R<"88\W.n:P馃`x-!⇦d9:̴S[%R$79Ϩ˄gHp&ug'Q iO **,/"WXC\5T"ݶ܄S$ZhDE]P * ({Iל3-^)o#79gmFwV-eDm<Ԙxyz\ 7[Fu(ѢR#Jno Nc`p3@3* hURRQMSd5`$Uȇ @S+ˌ!fL9T&uPB:kfF4D6'rib`y5&4`j|U&肎rYVXdlEhATt=E c,Ft):x̡c$24"!Gn?2H08W:#I$5.FѰP#0%Ӗ8*D&ؒLC 9OtIܷ.'qӽp?JXVNwU9Zhxp ߲o>?Wgv}6 a'_l& En:Ghxs; ʨ>~w9RQ$=빴9HWTNGMxNonUiY^pL_`ͧx- |]@;h靊3"%?۝8ڌ|oW.7sΓFR ۱{#&1I Ɠ" :I&YDr3:0JARdڌ@M*ъGSΟE1hfE,DTB& (PW| cKlįP3ܜ):TTá^@1\tl oݡN@T(ߛep+ UD> -Ju5T=PJLDpUY ԰dkF:ST}wBÄ]DR^mعo_mi_ќ_>d&zWbLLA%a mۻ{8p4mMkD:*NP6POE7KQX~2//.Uhr@ec;JB@RQi,R)J,~"1 Ey<|:@Ċqd|?h_A]vp(6 얭ZXlqyAGBw[cU%pSk3QxDd,J# ǤzoAҾ }[8U^$ ORn[iܻ[Pi6V~4i?9/{j|P XQ]gԽF۫qlxumoyGMhzFc&R;O!M/iѿj;UîD1op@Ko?D;y3fs{" -Hna&thϨ\῟us: 38}7>.!?{0hGd# 72)3!N\moz7KďC4L-?܌$F#㥼ZJηԮN)ZKѿl-]ֆyT i| UZEQ*iګ^*6;&j@B3ڈ-QvkekSRn@`.|U1hArY.]A3ͼ 4.2dDݦpw>rmBCZsޥf"}y"}T*hvL,qV%ZnۚۄT .hR$"PY $g.WVFmCNb8at{.FuueȲL$a:6aS{?(\ !(YwS;JLQaLo߰ }̎x==R |nxus2XFնL]xSxg-U'8q)6S\Vΐ:}Q-$]s{[v_ب%m#oWT}K4gbBe{%vY]Z~[0nI|8f^/ti7Z,QĹ!QMCl*݊*Z#(۽R4[OqLϔc)C _h Xa;2F(^@(#~;d'+"Ƙb;+!QVIm:jrNyN?VFލWUd&Cכ+yU|^1Z2˓r.O+~ AtKShFFiMP莀:jփ;望B@k]1E'?zٕprͰ=v;X'wx \!?S]|f %YiP0UXq^2ݶ8;EIw)щ`S@o NcAmc Vl |QkMs!3;G!Z.~Dcٍb}A$e;HvPTpwɩXbU$ *G^C2F$Ɍb林w91?Xbb#6Fwin]C_[m9v_fr/b]t("I1vNޝn擟'6{Gŵzo0[ȝ0iLL~w3J¾+0=<9>bgS[`r:P]/hLgǻ(fȭIρgVҹ1y C90) ~K7~cTJ|ws=闵fکjAg'o]]6Λrp)€Scd!ItNw׿Z&1䋷 !0'd^4 UD%\ 1F*!P*9@ <2.Br.1ǜtAUcQ>Zb>.NqjiU2%qbPPtĀ@Q_eH $)x ᲍17}lgLj=eLZE@pJ6؍(j (Ƥ ,Ը坜Q70A*7|ЩoNS^龯T}s}}*uGU'ws;u'ҹ-r0['ZljtW:[y[{IO\๻<)i>=|񍝏~.߼D?B7 *..(oFs0Z)o5{8vTq_սrd.(O~RتנY9jz(K)](E!m/yObwЬcW+)uyr&)ǂsS -klbz-ݤF:E`adV<=&=f M77͋'Yn(XUPp+\ڍtʹu9"UJ#O{4ڢiquo lMםC,f~O+(QRTNxVP%i6Jd%D)&Mq2վ>N/|8s_F]kp&N{f >q/{0RUn^8QxE*S.K# K^W]cQ"m&z\Ww7?gt;͠a_Dd%[y@9dlj  ^0N훬Ǿx1UX3Fv|=)f8UJ XGXXzbq Q%R|]& - {Gq{72fBT*{*Ff"n1q̑J<ԃB \t*{l\K:S`ZL?0o_<ﵯvhrIn^ <86{͡U-C3'S)q6ɨ j9NWôŮFs?͵3>㦸VLk!Yj_/,4_ҿY󕛝f 8y˟?>yk|0p?N=aLzr)sWMqֳ0%??y ML;7myq} #RTn|04P:\R3S~T1{!\?<_6҄ =OnBv^Az BR̾+ʅ6{lEE!Iq?Ɖۂ F!ń\u|+e[2s=~x{"67)}9.,R~pIY_%ZDf.4YY{Ok˔$0UM})p`އYqu( ʆ,-xKNξ}yJ72T+@u͐XCp! n4Zнy18[ /Bf눓 @w;DV]kww({Zȁ@?τ%: "1nʲ~k4dpJWZV`[t+YڲCԘŴn)V cjh1wbBT6t?6VNĄIVIJR֜(Aj2MŔTg# 6Q5mMי'N5z jM9i Oia:!Bqj|N n+`!(?7&G:.+N|W@[ z!>J`7uq׹v`X%˹v:䒒0,(;L*r>#`[~}@. L;ȬRHR,+&kp U5%fuő#)DR̔5I^#RJQKSI!*$FBX 7Cgm}9j5I2kbUIʚDsi˜ TTI,RqtZ7)8,0D |mź{LT)yLKN˔Lw9,wn۔b](7_j)/J@{Z'haoEa%3dL0LɇV RJCT.,9((H%vNW{T} ^j Hvêiz{5HA ٸw1dS= d_+9@eszds 2v2Ym.yJSj; Wp,` wbaލ# <^>'`ӽ#Y؟# Rݲ?Ct/zuwWIcqRl.]xַ~T6*?˽ ՙt6 627#wBXn)ĝɫUK^>;[ILj `?q|O1jP;T!,*3٩rgbsݾNբ]%飪c!9ýI.$Cɟ8N8vƂ2spXpZp;u \=H2_ٚwOLI.3J\ .!L sMqm;q}*'P1+ET7~^.*~2v:(,W1߭r~ 37 g>Ax44d*{N˞4L՛g4UR|PHo̓kU,46NUqVEIz=`wDynɦV)-%-RI!kO0F\!ukRU&hZhfG퀝D..8yA6%`:P-mG" tÊU,NfL% z*@XeimH*/5C ThK=HA$<kf?xC*Hg"os92 ,J+P e$/he1B a-H\=k>&XQϟ/Ah8Wx#$p;usy0TTs¶[Z(gx2-G9OL$4)rE=7 n. ylbcC Ѿqg;׶e(_1%5c`En&>,ǩ;p,޲{~߶ tr_ pMm4eYPV6RU%+- BDaD ^iF|r~D{ mԆ.GKJJa7[RͯIn|:p1}o7A8UUj4KR@jkNWkI9!@5wAiSQ~X}Ukb)p:qJ^,_j})Z-)-H{6[yOOb&kO/Uj+NA0c^^™B."e.3l=NPΣZ'([7s䷢CnҶTmw_?QEhY7Y=?:h$J%FXO|V*פCO%iL/&:%הɺ7{x[XX! ¼VMG[n9-vjϮ]eSEUoTYFoグkVnĤRcVǀ.̝c=:6⏻#:$킊kP5ot>Ask^OGM&^ -OD#pS Íg>w^g$_M |0k7_1(݈W|6[61R&Q"ïUoNjO_T1{㤼 WZU3Md uOCfcIvnnPE"lX(q饶ٟ\JixpOaNhF˧Ct@h8'2-[rA6cdc-I6fP! ^rH˽X,kd#0ON<qe b!*ƀCSEKz@ Re,_֎sDW O*:x!޳'!j2R}^X+W)֚LEKŪX@i٠EUѴkQRj_]-c'ש)4&)Pj %1r HLW1by_֚涑WK6yXܯSF`ˉyxy|X~X=0a0yakLÍ CeZ1ыy^g46nq2d D{6q%+m~%(cF&qEces,EiƑ!o8vTgjϰlۇ8enQb@BN G5=$5L5Ћ=wX IʨH%FV̚RlAEv99㊱LL"Ѡ"R]`z쑱 ZrBp:m,DeջXҒ)(lJ#u* NYsfQBẌIAXM}YLiA d7$w i3-rH$-GJ:l]2OAӑ!@'|7ΨLZxEāTS%a0DL-$!gԀi*2~S[Wb?o"xl{χ M3,`E}i=:||JF&X ++,]>\[C7W˵,*Rh`4EףO8*&j`o׹44)vSDC v5At Mi4kЇS1nE|_ m_ \C&z5-1^,} kF)!Pa,%Q:r-5eHH!W#]M A6")qA'*Ęм@`8_,hBP`X PBxxQ,AT*qHM,6..ZKʵZV=8 }٣ I dO`ph |mQ)ٮ ZDo6Wv4x?WoEɱ\ Jc2>Iq -1 \FUmX$Mؾ{9sLZ^Ƃg4#˙{XlĆY!o !vوOg?4s NNg;fgJ$c y/!!Q2r/71Yd+ ;SLs76'JP|9iϟ@ 7VV8ӽ]T7P;@~: kbwEN-Zr[X/GgѨ+FΛB%RaE"+bI/Vx)~qLkyI9gtd}b!E*|ߜmQ\S;[/ɕcꊘEmsKw"*Whz>¬wzh-C~t oO bv60rj' \w< Mq󇉺P"sUMmta ^.ۘ{k=Pcɕf2=Xpόu7]⚮yl|lb*{OՔ_#Fd*K&w 3u&GQR{|d%QwietP-x{fJT20O|X4,;mbK({%WE$Qm 3,RgmV}n]g?3ZCuFwp +VoKuݿ.?N'˦ud^HR?Td Քj]=F9NUm_r0||}VqZi¯E9s[tb=o==Hc';~na$biCŃYܧ yDт;nr=ݍ6.:$"aa݅12$zW0k؊6gI >>L3WOXo, uc9Bg&h?>h m IY<^оjNj4d7!P=j}{,x;.bNN$ k}zWnbMhMhLS)ƪ^En“G|c&?/vN$T&F;'[n9_T(Z}r%FO+p-k}|I<͒齓%da^kK¹-N_Vw0<. E,`^EbD?:{u݉6PhJ M\~eF֍($ㅆ|}/޴o7E6sJ0:1[[-XIM-ʦz\=vSQ^aT\$#mgM/Eɋnюy+n7^"'ז2*+2."Zee9v" -ɻ/*.2GhTU?~_Ȅh3<`eigl`Eq*P4ct-бΛjoi8_~+9Og*Qɋj1?3;Z9V]"'mZd!N ‹䇵9GqJRt"8Ôh l1sΦ8}hf }O~&wAU4% `3) LiPh5'cR0N`|Ia&e0ҌH)) > -d|I+ }KV].'yF 3v 8Cg#ǩH-g`#3/l)k܅`,?GXW#c/~VPnsNٻM|ϖnCWHs;Nl拈| sx >p\oÛ!afPwwe<܂εCUkrR& !!oPݙbXogD ¯y8N@S?ǻekVOޗLH@PZX 8N ;!xF9 LdN:s^¯7{T,TW8̉ZfCVPqi3w{I8 P&(C8tfcb PjLFQ0dJ  `& nl:BܰƤa<`a6Z3G%3v!:XKQg\heN;Sa`2 ^ q2ʑe)Xky+Ƒ aGʦ vĚqPx -𝁏 :c 0 1Zr0O 0嚂@lO $SH4hPQŁv3"`(%{3Zh0̃_[]VnA]La빦I]VB|9͈O3 L#2o7G%6XߢX?~f/0͞r.o|2WJ Ԍ+3op)ʤG IXЎJfʚ6_Ae;*>d,%55I(jb".v\ -Al ĪH$Awny68F=JG8]5zDrovI{o+ޤON$]&k̝RKmC3asrIͽ'2_B\pp™L 9ulبpZ:ɼ gsxZ93ۊ9GsLۊ9/-ŜSV\aui-.Z=&.nVwD9fqYz6\,m-. ^\,4o-.XbIbDv3X,=tH7z6U9as$k N:)Z ˣK%(os^ 9e@/ЬqxTg77r)S ެΣ}`a ),`ޭwb9Y!\ZY @y_cF gv}Nvfp8gSb7&+WA)cVn&@s扏Dc@|[Q$F_ , ;ĩdޥ9f9HSo+ q eHq.r3cBRCR@ 0g'sEV-7&l^NWg Aۣk SxՋG z^׽bչu\+݇^6?UlOBAmqlW#oit_U׽j6 ׼W(SS*NN%P{feqܽ^6/tX(|2Mq3iW4K+ִ{ oEݬ\9F#ogi6Mo  37=8jG30?_ΝZ0yF)$a“}W TkJvsϽV5TSи=W{o֘)ܟjMJ;D&LYDO_JDfl/?~O^C5"MT; 'oCs d9rx>1!Nub^fXx#uqBn ^| hQUe|Ǜ`3z%QWW\EStJyٖw11xTTQ1vn*nSnE[rMҩ}NLl{7Mꤎwk.)eZe}M-ywCCZ :U}Br˻ %»ꤎwk. 2M@ekUzl׈`VvZ!ڹ@5j䩝T ŻgVKX-@:%t=-ڙ@bzAZ Bɾɾ't}S; Vu5\%hOg]1i Ryz5f&XwVWcTWcfu5\%aӫ1sDWcj̵ZGj\q՘s> Du5\%"՘1w5Z-A ~z5fTw@Wc4WcDIsWc$M֘%;4]5f&kRJ՘s 5?B't5VԘj̊ۢ1ƬWcV՘k&Op1w5z-as 1+u@Kj̭1+u5f}Ј1 NƬㄮ܊&R^Y3cjZcjZpNWcnEY *NƬ!]1DEgv ЛEz~?L^݋̿H,*`/ooaџM@~~/dҷoyouйBPIϵT2ESR /f ?4 p-J7LZ8˜\ Z2mcD5KAYSX(J +'B,NHĭ XX -|n`rW'Gv,4I#P !> ql#sr8v!d qz`c" _N'pBl9{K1LJ("4 Mda){xaZu9֎K/891qXhYz6Գ@B 6|-Xb29pcS!%NR UE3yR;;5 {{3M կp3?/0[F0H;vܿe>\UX+კYDhg-g>,b`uv^c"Jju~hqQ{cT|XS P՟~>q)Lu1OO*>*3}b>ò?MqEc$dLqiZWx_B?'&'?_+ B4IIy3ɽXOrKcx5(aJ(hJDzܙyIHT`VIX9jRlLg%L" f1$y΄Q(˩jEV!ϰ|.cſ'w=+ΞvS r} hFor)2'DevOnB{Ln4aO,gG%wHF ɿ.>- >~ށgC[<2w7Aă o{p-~%/^F$@JaHj/{*qr[]EH(!5m$JP\yC=FY W{"7221}ϷCgO-.hHK7ԄJ`Vcp8GzT3@-ND7Bq΀uRY4 %ԊA"Xs0>#AVp52|&zQ i5B6\S@xc1C@wbOb!(^w>F{0to%)^ ZNoFfo`n\_l}u2Eز67 aէ dAϯ{ Vx  8zp#p,;} |Vz)?BO;#3_|F&U3Y3x G-J3k)Q 5:*:{.Y/*Ԁ$=AHפO*ROR1]W$\J9.}b5mP&(c9܏3O|&J1$b>h A@B<>d.}^Nb5שȯo^6׽iW`^a˙FVW!\e9ٟ6(ݯrM ok 󬏌dƌV~^C^́3b~ EDvc-W)! M0S]\ΌbR`V!?[0ii̦f62aA`>5AÛnI K:rNΛq I}+F@\)FG.]13m? D*z@l:HDXc"0&% !EkC 型§@ugT;*9e꿹:'˙tXP9QRָt bAS4}\y#=ETx-_$NJCB V}0%Kh]."`6?N|'ͬ$xtW nBf&d_7g=vu쥱>.e-]1yx;`J ?=y`nR})dKi{򛟯NCbeKyX,8ٍMk7 [%;!ne/~6 {MFo(9dٺTKz%nlS왾]Y@ijk~-b3 J^JXX.5Z"s\>hdӘ0]+sqZs,8Xs5iOm'0Zlxx7ɇӹ:lx\j7SoW"YKjӊ.~qs<~ܳhn8^D{_g7tLěK ̞"|?=Wj(j|dfgX|xpyXl~\;|8Y:,S{0'ҿZ,nĩ:z2ss0bӇ7E2oywtCĪbd<8~o,|٥LN*z7KawbvyN=B@'U(g;Xx!a1]ץu7pMbfLyI7bՈIZԴG]J{;/B]|\ujJoϲL&)*"'_|sրT83rӬYw{!d] vbn)b>ıd ?C9b‰FRBT Ney},E>SC:ĭ0"(KWZkeﱓ%U\p`&'!IFIK"jEL(4Rqlaa%(Pӄ)JHm8 %/-[wbNR׸b tsr/$ݹc܋0r7߀ޡ^$|=hđRqk2nNт(Ĉ! &26.Xp$ب"AgX`Nh(A uyPh.̧7nVpwB(_l.txr\PY;aH"hNⵉSU?DF׵Du˳%湷=z;1δ`VbLEM|d[2]!/mq> "]jϞFiX$N;/$MgθlSps/x+)} B]oC- c2둌-|o-Tݕ7(LEI/3"PJ!d<J1[žϥ*!iԴqH#EOݿӲV[9BÃT:}$鸚6ޜJ_?53R(9msW` *Ƕ`ºXAŋ66)JVHKi5v`"ă`3Β:C_}_ug~=cDhG{?54/q;:IY|Zi "WlvA8i°`z惺9İk-QWnwΰx ; nsa<3&P!HӔuwy]=nygeQ~󜋑#UW'c6fކ@dGu!0޴N}U.eW^.+ `ѱ?̠"vI6ip%o%Ǥ=xr]I_}2&-d`BkWe@kKRliAe~Hy{wH_|iiX5}{Ɏ3G0cy_7 򳫁s'HzI219!4׽Xҍl4e'p;3Cjige˜Jѱ?8z[->OJ u I*4plp酖\bQjR^K8 B$>.|nCYs?\9SHuu.6]E& >lt78״[8E:cp k:,mKQ:ݘ->D,׆HA/i u/꺚h],dJtD`& flÞ^.TXjv2O8^kc굱.4+P{adHoF{bɍaRKĬWXs ?]܆vs?xJE瘚[!Ʊ8sӮf5CdS48Fc48FрUEbUi@Ӓq*QmGGP9'AqpeJKc48/,EtP&Ѹb}z)urf&RpUw#d9sXյH/ ;ųw8CKIu TY@hCDYK{ 4ɁHVh {d'jqѫW^,DCcl0 IIP0@RZaV%teծ̆7#0/#-V' NQ2@WRP qykB\ 24֊j61t_ ^)A޽*K.]z@һzo]m .zWuMZ(̂q$ۨwˢ\F5tW.Qz8VEd4QxhЋ,}Wc",cO A :5"%);>Hs 'gc@%WEhTRu$8(VF`0Z#؎A9nߎ3*KX@hif\@*1rX`EB n XdLko,T뒂+#.<$UMl`/A#6e\%8{VFvDEjdV2gt鲤SIy|漴Of/[rEB7_KaդӊSѩUxeQ̸"h_b+ѪoFR1$[fb AQ0r#J)(@.ee`k:0F5D_N`3D|ZHЃ(/MePbɀ 8 F\ hXmT7>0f|]a@`UMC8C(}=j[nRQ+q4Fs";8RmK j"qP0.#X58q!nK8n=aZ8ė'JźBRUH@P(qFye1ݯ[c;[W:Qsc[39DxY?DT ,)ݞhqF1#(z2#Hb<#\FaQc)YwI9n#&@l< gܛkkivY%t{5_:x{_83~AwxH٩wԤg[ }Oihd{[~wP(!,˔X^J҆|܈#uQ>AG'qX`6OgX4spIA|?|Y*L(-d 1@kD۷#pK+ ľdN+0jtBz bh?.FZ|^J |_<, 3!)V_yyLë(W[kA 9cO/Z^\wG/71"-eY3hKeDR ":ށJ^H sxGށ7וC5$Qٱqʵ ؿɗ&m2-dm4IsEO-.-݅.|\}lvܨ/͠ӓ* ,GZw|9sH)b nB1 e@ʿ~=trT0)v҇<  fM]rt[[x\#;T=1;2I<|5: #؛aTiOIx״qҐ31Ip~PP$BOd!tg qSW+PԂXHg5Zt?~zhkDـp0^<Pyc cȂ`Q,_p\Oh|gIS@,xn?T+36귨'i&nYi>׃Wj8`'_wghN8e?= [}Emڒ]~[.~\ܴ\XO7!#N#Vi(9iUK܎G{'pGC=SӸ݅YT=5\woV^&v-xgTZfT/?LZ(w>|Xv155漫OOn"XmAa<yI1eE˭8ߘ_OS ~ k̯_#BHWN:ElXz9Kp Gd`zp%LՈ.aL&}_bt足OmW=Ԙa:Z08qZ,98R11GHIJ IJ 8q2dW'ZRZJhc0HXxR.$9+M0"8 QJlzR̆f6~SIۙ]6_5>f&ݸ1OPSĕ?ɏ nn~Or `ü[kYPNf-#b.^]ɑ3AO`@Ba\"uǑ`QkJtԂaS}yJә K|Cu j7DMcf#vy<` uLĄrQA<tIPi5SE"Ă$$%V.G95Xrֈ3u:f1^Z4A E|*jw\Ѡ|9}pK2,>\ߣag7yb6-?\*h%.>mc8w~fOgN+bֶ瞍}T͏ X_xzEX>MEuVRzgsW\iIB"zL1ЎH)SÞr^W eۧT&TT&).yqQLespiLGZr/g CED?dp>/MBm{jz=j{S{U3zK@KThd -%%#LQI—̴F6L|ca4F`~Iɱ͕)m 2)KZcb QQa1&8s^B Fز16t,D9ý uw`Jjhopv,pB.:e@˭SB'(İ6hh=3m_ %=\&mt,,J!}r6vb6E22 f- ձ7e$WlIVa-َ&$Z⯊"YՉ27OVVO`AM7]#-r;_GtݝJJܻ6t֗tS& U5ٻWjBV?wpi׉ /v3sYIHYq3 `)(iZǑ?ntk6$+uu ֚In;߁ڳ9UT{Ci$<VI_-yr%v PxPkX/@܆w1uk蟓Kwo^2-o[bfQ;>s &SD҆晩g6RB \ BkT7Qaw6 _nwIXyS>ܤ:j)urbp~V2kw3HVv'mxnIƘY/V2$= PN ]XV(xgЭ&oa'[V\|FlcK6j&^uV=az[B2 B*g}r !G^=+yGC G6QoVk:Az5INnOrr{ۓzr^ktJLe+4BR ō$LH8p"ƵS/ +L7hعrv:[}&u> 捶6Z[P1^a$Q?pi 'N%dK8[º:*#^J9ΫfD/y)'[Tv}Q'uR'YUs68KMFWPFòB@}j\4F|:YK:U5* Fʒ[+1T2~ :*UތGĘB-ADvqZӼeH؁->-Y 2Ug2/{6@C ѿj]mi1ɗ.!˷>j)k`.yעQ*ˡ͈h~Y~>(T8#?Wwus6V3E}#?>PT//Nȕۣj2`@%xtf,O N,&ȍқIS% q?PyZ#fsE" !*/s1Sa`IYa/H\CrQiCZ214r5ZKΑ&ZDmR0BoˍyYZc p J)  MWͷ{i6iNo2Γj%b:K$'bO\8Y,Zѥiq߾[є\^c?~ !S=Kr4ƿ)F\X h~"VNVoUf%l>FsH3j{ȃ?Ei⫹+j==_lu_L0y!`;2oNĪ t_..ddtW]Oo*C Ry_6miw&x{t7!2 m,@FCUN6$%=ѵU^l0} mv16հoukzM%>r7̀_`ϥ/Ԟ%~5[A]ۻQ e'3@F~+kwkOԪ F\xo?}AssXfz|hJ!41 N$G^qՉ2a/g)@pn9.)IajsW*]:CږTb|{Rt"ֲ`N)o_Hx뿖SҗƔEw~ G܇6x.Uޏz@n@}3K=0]%օX2- X(=T-֔}͇k2>\O&$þxfK5ϝ4doٺ/w ?_^ 1J_Z=jb:B  ,"͸FCuu%3aHIɾoX{T6,GvX{DU=UX |ݧ#F;G:Vj>#~ܔ Oi bUxtC&gخ/c%C|v)U.Ue-^*hOx 8UGYדC'[U/-4bç۫Yֳ$;:8rՍ[D( 5梦O_&cWw9B0ږ9Ek!y}"ld<a/ qzLp;ކ)H˹G#xH,R|>hvt1g!-{E)w -}= [zڤke\F*`a~|~HFp?ev|Z=֯{exz;&ӓiMu`}yP~tSk^y+bgehx4Z揼=,w.q|~h!}%!ޑVi~xt~{uKL7o-ߛޖ$;ւ ql<昚]]I10\`ՑQH~gJ왓}n9m.;A(Է$Vkņtw\Ϣ/׊ݿ4u?-ׄJ{X t(W!C ^uTdiN>s KO@Z:҇3a)v<|i(N:%n?ɒsɊcJ`B ׀ؽ!ďAu$^=}ZrrmA62e&)^M{ex5$W^MU.2heU)8 htEKT`,=N;A,@2R4GEw}Yص!OԜ`om0ֻ6ٻj҅V"{L"2L1Vrg E?H8E^E;JT….-є@h.47Enrf VKˠ3:("pQq0KLeBZ1Nsp삙ܪ1{J#$?,x(mILS@cbJ2wɓ@j%ֈ A!>]!Qq{״ƹ}#`a{-sj;d$3q8LԙX)}4A']0*,K{慢/٧N(6Nj^Jfcc/:y]_bqF›7h䗑'jO^v<*erR#K2j1$"!J,)k ;4BPl'V9Ifqt5IKMEI>2 ?&[n^l^ӣe뇾PA>/G9\O/ǵm?rwy"f|"-`=)/#>T߹tw^ܝ7?Sz>6bAFn e\)!Ps1`1Җ% N h|K{&.(e$ ((G qΛTB$(ݎ`fA듎V;JߤB%MJKcJFRlLy֢SdePZ(*)t.P7PܥBeE 9vD̐Ao]eu; 5wEP4sDVIT q1[HvMW.I^Xy⻟M)⹻qP=;(gƓ:MvA3P=K'o~vs?;f˞f7NOF֜'w9;(^,܌֟)0{K Mj(I\g@V mux b> SƠllH߄eq#`(l@$]>"Q5?%qb>X ېvSC]JAj+#Yr"L48n0伌I;gs[ ԛx!6sТV]* %}uyXeʅKўjg02- R \\z+QrFO_WFHd#XЯ痋wܓݭ6Xb젡0kf,^]ׁi$b}{]%ie|7Ξ0R`n 'm6|p& )vn]acLAPvCgqL8!LtPzT08,Ŭ $`d( ,LJ,Ktyѓ'h /R9U1 bFRZ B"KF 3(>GAnLU;UG-ʙd18f<yt~hmtcJD[N)HFz#fA̰teXԱʃ+H-D3%1lh1YeȮ٭@7aD1.@ohALbe dD "0QR|ý 22pvNS :0S8u}vXnCEvoYw4- Z@V7DB-5C.vv@s.`d4ЛCKiv⚠1DZkdLVف$Tٔ3P 'sRϊiLߋ[lݡiED1̣4zI=ծzIďOk@Y.B*; [Y, NxUz7OKJbZ}/B 6:ו*0uc- Y5Ex  ZwyxOcY;NeL^lJxy2C9EQq$";rw|ۻ͗i2E-^_5I#{9C١$1ES/]/]]ENh[uPV:Wip~][fZm gQFϢE=kh<,}FA0$ڦS8sf,0ŌA,(E; 4E O();WRh&%Ql"%-p(DY$V)OEmח 6 ˂^M&p$6Kܤ?.}֛ddZ(vU%ጭ˪,x9 -07PPhLPҰilsU=n<ϖ'ױ7(uI]˲i퐟l^qlZ/;@ &u<v&c gӸяLNG%"UHˈ@\vX\v3aLFuޓU]^lu8 u6R^$V *i\O?hK;)C](yQ 椊4Ɂ9-x^59IZv#kQ7Sl#o:Fyسͼ'fY}9liuZ0'7={'WAZ}r;ɵRڢQ§W?O6oS=ӧY=]o7ot}e{״ps58=dm8=y|4Y}gUdU(ǟ9Rp 0fRR )K{]ͧ}L㐆Xqs9AXq`u翮-٠z.MY4gѤE~4ͳW*r@=35 CEUKD,m g/:}ygGJ5-3sW3nۮ״#f[?ttqK2r88~2d4+yaybK}u}W?z{Cƪ`|P/ZW7Gm?ʾzjl_a)Y {3Qɚkt~c5]|xu]?"X&6;]WC̳*"R+\b׃OI溛o8fE$w.e h#fhdbiO~y??%'~J%- AID.~9\?/Ti*rvÛwx._6R!)Av72ߍ+5Vkk90lM~Hċ$*`XoeR^/󈈬n]sLܟOB듒by(FyW _nw3vu>mF)y_R _7mGdhk;^`7+6~Ƌf32 c%&B@̓UR:,"4%bRJ,P:8^K^'`ZY@߮c+n\i]o{. N6oSLǓϣ8pa9sxfvu/d Ǔ μ@c)َ.uMJ^AE 0^!cb NIUgf=(Z( H_I5}D|Q$A`-K[[}2$su72:|rE>+Jo⯧_\Zmʊ)sЯ "&ui=7P)ՄY΢yNDz}hW9$RȚW9z%zW}]}u3RTے;کBN2j]B e(|=$e,5;ʺ_J^ᮻ̦eze5+9 i=kC}Ze %?Vَ4v䞜9&QrW[~8CQkw@gUi)P+Sb PˠKAbŔpK"0,GTح,RYVVZ!9d)SkdM$hdb_i,]jg}%NJG;1 >RH@< @Ḭ ܜ>OЀ;# ڞwqW gJ|yt.kWsA ΢E58jpTU`m2JEJ:Q* DEQi]45gΗB/ʺ ҡ;e֘F˯,XK)FV8 +^[ l(9)NjKKD$xФalς$6}*LN}wƲ)[-CyL~FIJUOrCi52WHP_tpCE2'"%5Y\W\@zaj;.ߝVgS|]/(qʙFrv`QcK(\%]tk7u2D[FXPTBD Z6E-q&@7i0{k{^$-m%덠$ _%Q2@doIP[x_//%}y[{_jtwCVLm\k]ғ[$.=8 7/cO-pTPPe_Eί[W8Ӊ:XO:B6>]rޥ-bw&$6{ƭeol5Plգd4oz\t9x#-:1sq6MFBW@S(+&~/xB1>ֵ-nr͝ a įcOY=/} "łOLGsg[6e_"=[9c}&O_KWG gq4qX`h*=AetiK(p SSu /ڡ@]}+Aa.2肮;8+cY 5RrX@0@g4 m +Y`}Nlh*̾^Grt/>Uw8Vp|2*9 b0 .e`CmPܹp z3C^ :ZZs|.Cz|'*d˂3#bB6PpWX ![!k7B`sAN L+;Mg*RNdA~+cLJdN[f)KAbc \ Ke}UY_ԺBE/ȫrP2fŖwos|MZC@&9[vYqF3Bd\߆mi}iؘ9}= )6 v8 ]8*DF^Q`Hepʣ3WNzV_r=V}uU ,?n@3p*`qcQos-Y򼝄&j;^8Ç9ZP+X.oK'7WYRfq8)釯Yf!]ϟ* 6Y1ԨrgȑO`qƑ}l.?Lt:M 58b]2V(0Wy0H'-X1}%w 0Xcasvv-c󞫘߯shɏ\9zcO*<bn{f'"$¥W'aV>b8=u:]=͢Ґ'kb$r\Zk@੒Ha5Ü?ՄD3vu@w'Ϗ+j*hgf[&mE  ?{Ƒ?d7WG Isl*qFAH MSZvf lOO˓'Zb4BO˴BP̾^4fYaRmzpȱ*T 4ΧVVMsNe|llZ b)S") s*ytYY 9+B 帚9+(BfGOtٯR0GHPͪU@qҖ3&*FɹFcRTAˍϩF!4Bby|XE`FiۅKgYƘK%Usf`kl~I Zh8<-=R$LuS@V;AcFIc @ށCJ$ Yɑ'i6K''RǰIma xclc\qĂL+\Bfᘍm5)_cS/9!6a)hI,"V&qby1Tz1R\W&o"TjIRUF_RMcq z|KL,MJpK h*1M֖y3Ys RTQ׎hu20gԃqmPb8•Jb"IXrKHʡ(iS&}s@UM6*Bnr%y/.ZzҚ&ڠ0L-Mj©HcG9.2xjU\r}XѺ/dǓ.~6gEFnʺP h+<ˏTkLvRÿΏg.eWG& q+ n릘1cna7EƟ္1IE"g u*° N( X@-4㚹DzcHTUjahVj%q mHMzcQkN:Soy4nH;B %T聩O!5> Ik^mě"ճX8ʕ+Ccmw||hW$sO\PŴm˓vPhyTAa_ Q]w$Y!!F7ȝ y% !,ib X.&кT{&2w:ƌKx e^x9Ս &]5~Z`p/.^jRu UVjAX%ZZ\:D~^}v '%dnϵR4^+'.N}8H~D,ƞ؞TkIwrʃN1$YhNLbx<. !pqRtSEv0b8zfEI{- 6H3&7aDc#kS W _1coq0mx9Ӄ٫ksdE-( i[7~UV[7~Q<ݹS|&mхs祴 +ym09a\$mq3fb!+!A7[?M֠ ɿt.u~kS%@J0RST/[0\0j؀A.nڊ`l|r)*]UY3FYMjyݠu!l[| V*$37nuK䶷%L'/MaEOK_pf!\q dHp:Am$\ZMWu׌V^%,I /q,j9,L 9ib c3<_9Gg;w(0jYWcd?Noc'LҺ_bBKU!t܀Ş\`7`&mQFRTH2S3t|E\ }$$JGX#"j1*|u?+r4O:9[8iV8̙򈝿L?^nno7`{Ř$ _6c`4J t grHpGG Ӏ1Tu@f+Qge"V I6*LZUAX٫J(d2C}P抪`*1V.nOJP%òɧZO> :=@BtiF8ALA<{ΔƜiWoC5ZR_z]=UAHZx ׵3!{X'}bQn𭐍&,q0˯qNX`a˜R{أeU!N*:noʑB9_d*$}2JACvfMќNʌI qrdiawpMGh 2%=LZvRZvWBܫܒhű*ОURw*ˋ+]X'#qrX;ZњKyqE(hKRySQ')eXtJJmlG2؉e*%e_~3U%xdax|xT J]9zd~TRHTb/Y) vUM;ťj?,LF0+T@CQff#*7+eQw=_39kQ6kthm%TG4UqꭒQCe`끳L!'*X2oó?LGkh00ÆĂ^KC<~lx \&{-a`xZx;lo c7.n]3ʦNgw[olB?m7\]n7;kȼkn ;o~zyӯo{zMjzu@;v<~wa"+w;Vu;uw4t'g%GЯ{k{lz_7`9&*!-tOݾ~ߌOqzF ohTgqm̆޸Mb,~!2Nbxx6as3|7pۛ6<#n?U 'c86t ꞇgL} oX~]f?ʰq]ٵl8儌~}Z|~u07!v.˯Ǚ~X~ ˻Wӧw]=y?+QP;{9?{ "q4;=Aem&类y]?:msg r^w>^^ 2 >n{Y|S{g|(quwYps{Wn]喇6HOxнa '?x{do;o~~f7L}&gN#0LDnzOW쌜Mś$:,# x=lmOnp|Kә:9˿us KvݳLwҺ_fDZq;D u'/f8>;dF/;٦~/#d1=#,= ]v~o.&y7-X0׶4[UkU.cݨzVzKwbN,ɗ^6VQ8^`(f3adJc%XMY5[g]rr>뒃 XRAO`b{ɔ)lFGݓx5 "s#+s#;scuFM9s$n.Z _\Qu+܇uR'`(288|4/FXR Vp0U$!T'taQvR%= ?] a" FQ((FDx*s 4548^^^!?Cgh au~厁3,һ?PhtWHrvqHN< ic,( bDm`jS$N,H҄8 vHsYd c`e/!rJ$0XM }qFFTQ v%c;7RhSCWLr8YbDpo)N5A_ zQ딸Qՠ}5k`{a!I<~HC?~X:\`WѶWxk>ܞM!ʪ[Ӱs&xq P/H1Q8R$}XrejSA8$0 cαQ')kkpA ^.8>4 ₭ׯ`ZtvtP8^|0s'5Dx]I5:|Pg]!E8ՖHN߯(QF,pyjvazY&[e#.GVS#\Fفf!Xv͗<'jh)bޫ8R%!M{e/IWVkƈ'w(n T$vP> I8*4U Y)b^NUGCO~駃Ծ}ڮBl5o [?6k?6[g9CyEnC:. WcZ!ʦ)$p寤HQWD0@%kބ 1Ω>~紕Ngy+.if>>kû6>kû}؁ۣ|:!Q pʃU"w"HEpTF y. F7Uv!6I[S {`j,srAvggC5 ^6]1f`+t+!p wp wHw1oC~,OHThĿ[W[Lo_:DY[ 3Ӭ% Qh-l*H99dR8\[TtTQ/>'#Z T: {U0&8"hY, AX,d%9AvBR8%\Rz+y\.^g*]#5{A\@b ]Lw HRlZQ T;8%4-lՅ,GNZM(0QW0$FolИfIA-#Ώ րq?n;Aug[miNexwnK'۽>p+G ,n䮖*DP$^ϧRo|ʌQ-b;0\$DUV (5`=r>Uۉ+Ⱦ_+ Z k>X΁= Y*M \])ھA1M:P@,?S/*/?ӱéx|TۘRrVM䶓lJN\>ɑGokjO=lYyzh( Rs4 1-\cVI 4- v@UZ:P5;CkdǺ Q HHXӌa $<>Dz|[]z2hu=H0Va>v=d@ɰ;=BWPyncrwۦ!ӕgȝ2{$`Єu; GhgM]OI:g@LILRn ǒ.}KR"ۡ(!YrЫ$P{ڑZnJRXr$帨gD)wКB5DAW 1WxTʰ;؝;_loI+GIXuZ?;rSh#O2֚5#H?DIb/O=u n"ɿ_ѫye:Z Ce^ [pⵘM:kgE ƻ\5Vy $/9q5Ճ1\%i4u~x4u:sN\nsywR{ I9:vp>{lևdW*p4JcN5YGR1l]ųZnPSOi7i 5HCPtl4)5NeqBAxz;#z9-40z'~>gk\ NZmX OM6ϥ{U>Ѓ]k){W:*lbZ/8tؚٯ^&^|{C*zR hQ"[ۙ{nZ8YU8c:4kjsYZl4ꔡ$]B3O&ZvYV'N c$hq;Q-LU8S>UWg'ST3ULU8SVl24kmNYX3==(l\ڜ*Z O$h9EG}ބW>DR9{⢳+ CE^>S%@f %L hb@Ā^{Ā&41M cm>m S_>hq}Fo$0 wK{;Z𤿀b[zu{[ltH,#pS B32r5$U.>=h]hS%)lS]tƈh L9)gO9_ŔrSbYaEt8UL{Ň |huvu,̱!&*ctN?yO]*{*q/91]kW= (dڙͅ@|} xִdgwm!;C;ŭ:ϽkSgs r|hE/Eߝ={?˙?%\XLR9\]ٿɢC3*K.:8z~VU"y!8.EGQ*X b!W:Lł*pe`p̒pgO[~vp>ǖ \vNE=fI޺m؋-BJ7ޙTv>ԎTW!$SBQȦt'8;qZBAqF8^}D؛/imFN^떑\f%׼b3HNk[Mg):; 9v x،FrqIkWM*zHq1(UMPAQQclimfaX]}6>\׷#U1$W];.udXO*g6@h RI@JuAEȱ[5ff6G, f: Q6S7goW̆;.&;_\C?@r8@:̂XS- r*pTUŤÜ 3v͆.ojΏZ<<:cuVC+T(iP;!uP%h Aet&Ϲ*,"( Eܘ@&'غ(l "V 7Y~X d\!b%[ R^uYˆ@{)EWTL<%Nb~)bSԙD2PqpҧM"4yhǘMu7=bYF(68ϯGDqrv|k٬rs`|c]:Bb yE( B}JB M)Z bd61jfሁ܃l(뒐gWO~>u~\nCqK0Ϟ9]ɼ9fZ3+x݋ul |rz;9Ke\ӯ z 3Co퉛Qy-u0V"~,rèxUVZN7e[Z{`(=nr\cƂpd|2efJR 5D(Gٮ&4"}-ߏQnF oIod^6)+0uJiOmRs]+](uK!6iQ@D5x2?{בʀ/I8TwW ЃamAUM49 S5!G3gss,Glؒ>]U_ץ.7&m[S+GG` 1&mKv2G%Ѧ) @JoewK) I$Z}@iH*ѦԜ&"Wm]R+ Z nLx=)0g=c8E7ӛo Hl1@ Gp)IyS *Q[4`rCȠD uoTny۲Z;?">F0v?wd?xLW0xJ` QE( EeQ9c򑷇k$?P0s(`K+1&2A$ՠHZCn#c Ɠ\wc >` ccLH>mM_,H3Ar[{$Ւwc{CH'{D9  R-vC)9WCi2PaIiQ"dmX䖷-kֵ5zDAx=2Ԫ%I^+-;y3e۫I IP PQ:a!$-KMz'%iid-܏=\KC:1\#O#[uKo-AO0DcWIwd!z Iv!;U@-9!r>&@E1]0)~## #RQLXSLDA@&HuIhP}afH>7#RNp5j3M_Gt ֙\,#g cTCu-bJp=,$@ѱ٪HFMs$‹0Cbl_F ,S`yG fL7V*2J·&h E2R BvJS&g ٙ(Օ血*|QRSU>%K 1EkP&bxf0J=*KYB ueCB1y_:F'V6(o4vo@ 7rdR˞|*_K 3I$)\ C*2bn, i=.`ALM+Mu׫x$acE ĚL>,y"Nٍt'u-DMU, j#Sr,:{YwU;gQ@1s/GD:~Ms~$+?KY28u(^7Rf JftmJm;eY-2褣Sc =^wD+q1NBQT/($EBNd:b|AoyAބ%m( x 8Þ/B=񚊄@ 3T `Ds[8wKBP#B~ OhʖW>+_J ېw]!W!oWiEBږ5bMp@ I9oQx{DIVL5BЎ1-yv{Tִ(}S8{Z!9<-dOm[xa A3>-*RsS$QFMR{ v--]K. UL҃+?;eoLny۶&.2`j&Iϥ$D퀤?NL ux$咢Te$lo{h׏IHT8 E Wm˚ Hbk6 =ߴ>:Жt9#0*<H>(כ3'VoΆ,GGrfRӇe1H^_H>mM+CIνS@r؍D‘29 JbkX1-b~V,A6T}$r h1)ޠeMٍڌуr.=A=ġR[wGQI`fA0\V=!78%R_wwJOU+,$ .%-ot)h3Cz; Q8QA{={vʶ[ Gx}h QSD!@#}\P [>Im[?]\H9fMt!&xr> 0 jG,˿w7;9VL%rtR;nưY:=߄foYowr]tҴQsU~fSwz #Q)~uqDz W9^a"U^.Z6=7# ⪼_\_I%:#IlRR|?C< 7k^;}Md~~\curx<8g{~/b}cTE*l=/?hz9@&ǣ2:!9^+a&99TgsaQ:H6b.؎_lSq,<̧c'͘7ZszSv>dG|0ɓ݌zi@ ZƸ(p)g5NV <8i.?S3S~ 񏘐9#@n"n:3:J*uLdLoTh,RԪ[Pvc96+IvxgI(4 }o>H#w}Qs +2PpN1I OFn@7KHk~Z7ʹrO|BαbOԮAMM w I%*!fJ%tfv21`U!FGT J"J $J˷3 `ٮ6~3AtHT6(έ!&L8C.Hmd{֮Ur臻 qiEw!BMjlQ۞Bꉳ-k*ڣGҒIPsn˓yOvQv)xf#zI .)P> m|u9W_W(b%xe% _Hrl RD9?B8"q-c?4{]!>G'mz9GF&kHeF3_g *E,9*GMQO<&QjxI"ԁ}SwAِ.IXN ])IIA9 )Y @*hUTO#X=MaZ{2/:7,Ʊ$= 8pj)K52atTU YdI'J*4&J%WX0*]ɡGrW͐z[t\bf뒇#I7Oeg!;;/ͣ:=W YK]6=-r@t5DO6$xqhqM_,,4X \֋c/IB[">tNK/C=Fz"r۶5#:"8ZNy*LYB;9UA:[ϗF %v=_kT"JǙ [Bky7k7zmާWhkR摉%7,Y?K#Mh O7|a]/Gye :<@fږ5 "tYk9pi1i|*]$EMjk1E֚EJ!GKàe5Ynܗ^RI-}XvSVfOН@.rb` Z!M,vvL%O)P~b^W;"l; #]wMp[r%'J:}mۚ=c-Kt, oOCgmlyUsG!3px=H!lz >T)j#6p$p H5]2?O2%q7nGB#2)v-\"+똍USZOЇ8z6Uwq);eQ7|.toW9jUnIp$@>nzC&7zJVlUx[nm(~&a'dXŻlu..o~uszyoeju⫯rKӖ.⫊՛^,E\ßx;*i?RJ>$7׵c n#{oujٖyeωfwy l\mBЀq)KFڳ'Ir˳֕6rwIxٓE+%m5Dzիlކ_ocS\gaڣ;N$"}֭놔|F#U<ڝDp t]j)j* '/lOEX~^.tlLPs$0lsNIEkvQ[s;Lݻ4bwڨc}">a[!0^߾ytj^CkUJ6!qQUc=69aLThPVx|SC,# 15GQiSmM5&$(|ކn.'j dt0QfK#=7MZ5U(6IFLŽ|r:f1OLv"$UpB&$3)$AC~ܙigo2b)gl_@u3ycO&0Y͋wvzqNӘ/(wv>tKŅ;j_j7YhdgzxtB>otwS q.6C[ͺoqlXS/JE|pz~<6}gݷ۴N5ƄݪT"ruO19cr|tNXjפ*45 *{P)%I :չ6alkm`(vQ+P̴{[l-{@*vǏv2wf}I;l*eNbG&76-&:a3evI;Q8ϱHHI.4HYbJ+12^=)G~x<= c<8P Xd^?TC )yd{ҧ"෷dYѶ]W:9-Nwyp0z]8  Q X`䌱׋dPǽ'CL'jDEh:qq5TӚ\rgE'}TE1nN!q)9#,ZwXG6P(Xi/17\)Zm2*! Ɵ>}MK Μ_/-ɍ|a@/lfeA2+w}ȸ䡦* +*2Bx$֡Q䌱On#ч.CՇS1TMr'fZEE(KyDNA#qdb2ES%耻B0Tղb,@D1 \B^9T_dQ",!HߢލnfZK*Tz)6x0j# O,铜 Nkl) T>}nnGl}JI $ e 'HK"l"tEZ()Ms}فJEؗXq ;3$3+pX[JHB HLic>3?4 c "X?٘Y {- bzgq?;'rh%X>fĜT[D@5˚&7 FTVHƙoLjXˍo ildWƓ[[“8Ɓ-pVpSxy|{`^mĪxlevWG H6-{[3Mɚ+٠ @5Ki9 yAjDzie*O5S$i'U%y8'ӷ;́ي54Np f3=vZ|飑ݸSolJM))7ٔlE[>\DX',$1:XI 1DT$,UTPk*̢b>e*g௞ IeRKDT)2`wm pɃwP>aɑ nÑd()ciSp%6JWZ)d$|>Ḗw7`t!|f[K)C,zwa BWR@Kx:*6МC&HBJ,}2 Ԁk"NŌJƁqPa*b26gG?w9͜]]50t^{=w9c:kiGO#Oouf(L'p<-} Pȶ$JO ˩- $8U;EUތ d^:T _vqD08{K}$𣆃ĒʧIc7d)@)7='6r."2CM$ݟ[Hڬ+\K2_V&8<>}L! u@pJc( av7̧1v(fLasR1UqqȩXK$w[ ܮ Umds"%n-0 n@AhM GmItaz8W)[* PS.`w`fev:zujX5~@nKm^il*`/3iO.!%\nn`Fpetp+``KA䊽y^2vTV6x.Rb&в^R!MIn*?[OR!rFJ@SZJ͌ q]ZN2bc(bQ w L2BNr~Z)i)O8P2DD!қ̚*Wr0stˆ 'sh: ȨE'>hA ?7:^Qw={=] iOז,?Ч/ _[.؊4-P,uZ"&燬9Eع~3śrt{.rcስ UKHevrK˅6h1=wHEd#T"( K Ũhj^Y1abq[BnF(TxXL 0=a8rWk(I hbApE81&Ajm2^V|qEkr,@xgkC_PN)DfRdSKV_{,k0t7dcqn}2^d93bn ѠL1_6d_ŬwVb?l8!z2KVzZ׋-(u< ԛeubT<`4=L]ŒA"(AEL@)Z*$Xԗ"3g|[=7(x$#%eh9 8Q+ǂA?FG˙e~JhnL] qblScH8r> ^ λš(oV{J/OnJ׾< ktN]uRGmSGPg!$8c$4w?afnyC]աS[KgÅݏ*qRX9cN} Zq=F>OèY8@.p#jQ.ײѡݡ H][[A1%Gz? w㞧jvRyg~SiDRZ}]S |M)~.\C\)s .WwŐ{Uq p&~Gz?٘j[BHH>=)Hc{=0S# fK?H҄xϽU4QD}ֶ!PM:;p+mǨeVXB)},0^ebQ!b Y'b٬y6vܒJQ鯪@ţb%gN]m++\s_e&Fv]ebURK"}"! U4}WKaPeʴ Ugݺ*@G.1̧kT*f^6vqxjѡ*-l5 mX%[4ɭMZA-QVuvt6.3>w0B1Y%BIwpbo<Qwjx v?#~[ 9c&';Q*ɠ+M o;S) o '}1eq=WTo9WқIG `ˠ:sȂV6 r{kg}*j+&GcfD[._ЏX`~lh>{5,mt[3*뭜 W;՗*JD{ M q#s]JPa}X2758/-[SC[;Up\C`qт<]Fٷx^ke#{/? =˃76XBё+$N{Ӛ! q cqgi-F7%FӀI46*'jF)C3{6a54[?,"')@^ĵLev)-1~|~VbK;5Nغ7z6{vvx7޼'6o0<6?ӻ/Ⱦm^د|߽,4Cm#͍.,s6~X̦`u0- ˉjfpoȹ\00dcE $p9||X{]7x?^3U\|?_3$8߿ a~'ˉ5]E|ŖeYnI;Lf};SOtpk^)h#|'A$3*~Onj<>=|d8. /&yd:6H,zg'8}wC #='3? g̀?^G[ٷ-Y{v{8z p&L@K ,adboj~!([5 1Ab$g&~5fI^O/|.VomGǡ%_!˟ӫow==_,1\ \~{q~z-<˂i3|Z ?;Zli+ezjfHW7d>kVURGl"WG{,ߥ2%n+'<$]QzhA0@θOC}:. ԈB5<$Tћ}R9yܛ2LmA @]h¸׼ _>5\onVW։!]mzr^Kԝûë cuso륜FU~̿dwOB[EV+kb 魝޺Zo0UPGEQA[%,4 ^Q CKPeZ!C->E!jDd='ÒB㎸\;u$$l'vk'.R^)~W$ZlIpTָTY-BHɟ(ù //ֻɛk8_ ;%Ϗye֡ JY1Q^o)+/0¢BABE(z^ tCY<4 >;d6113}mbT1P_?V>FiTee9M[Y9YΒU-?:gZ62'dܵlz`W;=WKP::!?9 KʸǐKU?!.oeW'4mJ|;IS@zv~K>+xh\ǽFw;?y/-3:o.ͩ 7W[=De=DlE~ Gy^/i>9tn^5Vrkpqʱ.C[ʐV}U{'C4'CnZː&wjܽ F)FnU Ith;kҼ[sl5)!8kKI=Z&1{$!@+B$ (e&J邥`Jǵ$DRnѸ7mREE|QEE®uZ$Z~G Ӛy5끐) 5nĨ@5T )KVL6X pN8RP65Ilhk] d] _ҹ=jb|s|5?<_Xb=;{|oG!r[s7w,$H$M@#! 3;OV[=0H%6Y:}YN^o:ѡ*ѡ*ѡ*ѡZm3r]in̨60t2 *K[ cJ]P)Q@%"qZbH!$gd@3%i`) xL dhv [ibRU+UeRU+UەuO't{]p9 JKp"YY4L.j ' KsH,{CQM4G킃' iU<NHG GgP?n˃d?4dDl꽳<3QE#ҕYWB'VA+c<²4^ANEMR!KOL(W*'B6?b8 0v+vN6$)}x^0XEw0DAD%lq0P`h`u6-GoLsUaqHEChj3>HbL%*[nUʴOU}i6u(邜bJ^bm=+:TS*qɂ|⟯O&L\Astfֲ#۫ey%U\:{coغ$Z?i{4B4P>PI vL+GV矣IQKRj鴣E!YӐ"1^eA] 4ܵ4ry;;fpPV{d/ u֭R$9A-xQPI-Ai]t LRHttf=&57jjب9rz-FAĬ''ϳ(S$fa06/k:'f3**VzG 1FpKG͛1 uh^@LF *C 0lN&dX.*wGkכ_t8OE# f6>Yxٶnkط&oZ6& 9crM ICh] ȅ! x.ͩzc٤LNKy:ɻ;=WקuP 9|r2?|>o5O·w؛I|z)c|aZNS>5z:Ε?_,Y;3ֿgq?`FȾ5giD!\9g\%d!$d³VVh ;oyQ7ivr8oAkm9ک㪐8 &s^[pH"P둣WIw2;P /d̽#C=kafʲDV3[s&q*JHDJf$6lYH@ I%$j˃׃1P&ϙwhD1g-,댶-'e)X DRze̦2mr!}a3]lHj2q Pxb**4.)m3$ҳHC`.W]xH Pn6I)RU6/B.la]upyJ΁UG sП+?oBw*ck-yo"h}4ߘ w7ycCԖ$r{΢sQ^bdq8=D';Geċh7]yǎE Tغ3j=$gRl~,GS|ы+u9ZAL|E$H?kEV3]@P,ej߾䟗eԑ>WNC11Q8]DnwŨ TΣZjDä9g:54}0j>f2;_|npnb1`9e.ǚ2/hJw.rҖ#L- R <$71il4HKIJ9lef=Ĕi RUQ1Uk|?ܐ}s5K }LQv[7 y%a$qa1ja9fcMCD#߻MC?&!0_f<#pA-(ؖ\z=Zuq-A o4J !XLJ<k?0,ӈ<5F0e6fGڐk!Gp>4kƣab#+_3♣OKty5#~!u5˺ٕm4<ںIUU{j!w*牿_G4-HCdn'd૗RWr'f{Z'1⣋ySLϮg}Ԁbm9 .ń9ĉ LD>!#Kv+wSfr towKi%萜8A&qPٻd\=Zk{CPHj{ϲɴ~ ݯk^ bVbh&En6mœi~\5+a؝TwyA1K0 RpKX4h[!Y bV31zqd'WL?߯mo2ZIZþ_5[voʤp%vq5 ykeȯ0hqPr6C\o5Pc^ؗqP;\BCR⇸*{T c;R; WM^xpj4<{-q |QWT5ڦәL?_o }W44c_(d;E];{eHN5萆H$RS;,,-5PuG;dSG%݇yXCrCu2;t\`$]l/`駒=ԘT㬨t???Drqo%?44q6{;ͬa>?_8~W?yw}8#ؿ:f~Ŀ@oTZcNQ0 `.N(-W4gj+JOEZg  xBB1^RioL񢨊ƹf+ۺ Ebpya5Y;?@ r @$y%,KHj4X&fP]N۱vRd>~l0 oogq:ױP Z PλN\c'Gߝ~>|㯎wYٝҕ=F"u6$h%diVNc}3ջw:uPN&r4R)QQ[g#z`li?r4um)n-vOfgrl0\wAm6ֲ (tW$c#spi$:'cl 0@Q0㞗)xڱyU&`,b #gua{G^oK`Q D,,@ <-LchG :^po B xZ>^\}RZGjp3^@|~I|Мڜ  RӬV#eS#RhAkgP'\"i,u\HrQ0k<"PdT\o{CbK>s)-.HL^"1ÙdW%(arSr9 B@W r^('=Z# O)#md@'s@ 00kk:^vB^9x8i\Yزt{% sepQH*JpɜLzN8#VBt >s/؅-$CQj)GK!JI!LQuh*Je=i @*ʂl _Cs:I>JCA2&).RA !+K"Ge1%O6,4t@47ʈ/> \i ]\ÿûTӢ?~\঺{r0ү~<ԋk =abR?-oUmU~'cog V^^I6$jDO]Npݬ860`V.f(kw=zg<2(IA Zғ3Ky>Rse׭mqA;o,h|wymS&tk./fa 9ݝz,.J36fvia^Zk@uqni\fh#v곹?|n%<=T`X$eIl\W$Q2-wl:loF@[!pomɃ'Xj {U0/^1TGuܪS?"8R ə*J(bEѻŢ _&4ӹp4#A_: gI Z)++R?ڝs3N} \f~(eOhJG Gq:}w<.7WӔ_&f;ii{S@se/o0R]ϘK^my7Տg71.4#%;XUfC1InIW@>~ܥas̩/?Ym(F1SL ] fX-n}F}Fv0zhv!sʦ^2>N~Q_k1M̸ACbBpv)d2xWrJ>`(b*tɂ5 4jl0,4W;Vԟk2+&uD"S#GK RXS J0TB/}F\[nOϠTJ ^H@z#(r\^OvrOЊs=&լfFNs*I%h Z\g_⠹1-~A C垆)Nn CeibX]r+WDŅGl!Q^ 嵛;~Mߜ7GInJVݛųT|S/sU//3_^6^ńh5(<G*m J+Kaԥ˄#C[@~o.h!jrA\qބDg EP)/H%`K 5c"cZ}?6uv֣q>y䂬&ۍ\LIAf3֢^;0BQœ1D-bSi4&&RI Qvv)1ArOMSeKmb{5sbAq{y촹wGw0Z1:rٛ`~RSfIdPcgP/v#u>6t:߃-[/"6~bM_s: .~!aX_,^RcFVhuu+XM i@k-xaF§&J6&\ݠ&z+5C#Vlƞl2B}^@)P~y>㫟ayCb׃%&c5ļf![ivE%Ej'9'^܉ZqITWNd+&hACp*oMN;-ҰrИǩ8 ( D06(6ֈ(~k뽌 >SnmfHLwǸd_ŠJfY()ͭf~6YpSgsr[id;6`եBS)y1MGOBGB`-p-x ҆*2ID2bgڙ[l1=1$Y^8υ-ZKj+[~WW:)Hť^j_`⦻!(JK U q:*r-},+qxE-h̡bh2amS-.fF8b߀ӛq) H)+n쵯W/ʘ+$lTcF2 9Nu$Ę|TX$vT[҂1Fy>Bˋ;%s/*dwd}KRQETC* H 7R~0Fw֑Oyb̥#L,huL|>&mLgDPYA⟏&cN0 <u׼X<]Էˢ,zZM>GiLS &dJ0֩ l&Lt"'n8MYz!Ȭu4f D1"+p ""#Ƅd&Q*xHJHX:} 0:=?\LoBұ?QLtdV6KS(sNICS~"B.keisfLtA.a, n}u~r5nPm5YM5&WŏOg6? Eʭ^u+.z2pe#dEbl%1l(RP8 KX'1ib]ƸXJ${3Fܠ?nsVyUկ;M0/z[+Dšg}14'_) Q[Iۗ=EHRSŕɭ2uW&xeRa]i/DMXYpiC9chXLj4 甯םם`Y{֓j.M_ڭnvۢV"B9jf|n"HQJ:GW^+fW :^] X/ y! .W_/k0_:RR;aE7ÕW2Dֈ UԜWkM(%7g'چnUzAW:w}ĸd9!7&RXDaҐE{HOp 9yVj(T.x~م7M4$~_#7K1ˮrEz??WnswWNR`yӯ_n0I:pHH>NG2Љ7*2#^+j(viّ;;*]#K Zh*wH2Blh57*t9=\/ٲ_) ){_;tjpqu+)w5"ZJ5E7\Rү3豼tNX`MkI m3JBB2[(='6yH>Iޅj4Z #aZ:Gzk*n3khk}&}a(|`0=S-o u?$I 9Nrn ۰L=]h1*+(D?+H^bɛ>$=$GyaYOe&ﮂs ޏ%o<>Gah~FOdndsNeKܶ5ϕ;d쒵%*S C3)HӄF8i"2api2KT RG.E1Q5\JP b~ &i,#,aq ,(5 YKoqWqGIL ̆qLA”bZ)M(ML&KId$6qB,4V,Xo:J I FaWFR.pM m0,m$RFpk1cǤtD4wX`%OUS%HXcZaEJ"l:xumolU%g!#m%H(6REqk.I r kE[؁QAly['_oӼ r>wɄat`rf̰ý"p %o,>fCvCG -p "BDb}?#/׃ N9W?e(^ Sfn?9<́i<[hs5 o jзbV/0n )r)Շm2ՄG0h]&ps ?FDpHjZ!05s=,ɽʣ2#[t PkB0q=M¸ _̚[{1֨]Qc6&Qá5ϻ^m3ŬC.'~c ܢC#P'{W~\VxYc|ghr]&\,9戞58?ȶ7y#_ztr3ASA5iĨ1OI#ʥA*5 W4"8BQjB aAժCׂZ\bRE ک;եtS.(ݴ)%R"ie;u 昘(03!HBY4. KU.6Ͽ\0EsueY9$hZTq Qm:6%2RR0V*4Qk9bW%c EaF@„RŚwU;$muInYyl \mڼp2+mZpܟPB|ʅɉ-'r}?]g$5h ^莳83V.Yid\;; e:)/y172>L <Ʌv+$c 6Kjqb#7s_ţ Yn˘Rx-Q%qBkb%w_e~(fBG62'_`re],f6\;}(4bŮ+f ]@#?-'k~w3&ج0S|ϵ!MbP>K+Q Wnl Lpӵ.\EB)豬R.;b s i~Pgt2gKӛ֟KTe(\uzay^6Dh!bJx "NpX7uX\jZ|YLk,ߵފY3+UlljY&twݩkQj ^9\^bK߳:qJ4r!2+봪GQfKs" B2d# 0iE&CJqC5aK'9RҐt֧8g*ewH-Q%V q9XYݖ˃YYWnNke5n;μۅO4yv );Yiqꧤq҇KF&XQ-rZL!N1.ӠV)>ݜV/~9J7YVttSR\ ѳ)DDm6]t[_lSz -wW.'Ԉǯe.Z.P#<5$P9e+eE AAxMAAFu{b@p0Bf bY1&/ 8?(T`.:|t 9G4~\~<kA7C͟{n&[[rɍnۻwq:k⽾һ[Kb=.lۇ镹6C=gXW}<,)d/{,[-],lv=0,L^z %}0cs?xʰ.t ɠ?FA_7w_Ҡƣ0H0%lR?t|,߄OtP׾i"&v/H۰D4P%t+Q%"]om+g+ KZljq((񲾷H7qdnn.*ng Ꜧxws5I3ǎ9{ sNzO bw_.S * U5%bV]]cVϷEKf}gin O:/ fe/O$S֯gAga/;+J5JKD+M Ewl PhH+OIk ħH\ԯiI_Blɲ[1|:qR :0}ૻzB  .4+ڽf&EUp855%$9*G(58<"IJe/“>RpAl`AѨ5#-2h#K%XuFhw)b=wòB ԟğ( ߔwtmD5IV vPuzj(TR4 kMKHENGNCDž x4m\|Q(Q*qqȂkz/hи~\=ֽg#K>U^^|~$K @8OqU7Am_;~%C^wnXpnsOӷZ09ݽT2ZRJz6bӏ\0K(#_6”mS릈mĪǔv$$=g*33>ңWfwP2W1|)^Ŝ'*v<<~O > mggoz*giGCiat_ ފD?WWnf'ԣh-gN^k@ Z(_¦&N+ J)+qzՙC69o; @LZk)o?.s8j|w~' ҒfZSzd)4^e64کSҏUm4VLSm4VLMI;7 %c %"DOic!*eL{gG鋊~0BrpGS8iV B)Wypn98A/אm9wuU-Xm9hnyw-9r0ȏd>k9I59As.~蜯7k^4P ٗ?dB 9AXoh%Ei\fʻuոy}w.Ng]DaV!Ԟ;R\vDg: $\(6Ys6^{7b Q~dKtzt'KTHPh),,J )F#BXV-X9hYd5/Iar! N8!`ߌU\{#aVJ;Cbo|t^Bңҿv;NrE v'it"hSx8Ψ/C#0o}u~z' O%DxTbY9oY0)&933Ԣ0JG~GcDG7xJDۮW2{dh9S C; r(""Ւ'NxȨq}b1Bq5|GaAnT JpS]gy n*e]͚I8yc}YaiO97[KM`"ΣB2Yn%ޏu'WQ"^W',1oZD'@^hyS %;g~ٹQPn z3BZL#{Ol[Vk$ 3E)+)&yai".X T)X+dZ:^ɢDנ -: GKe;GqǗi&7W({&$ QLS_'ps]5($B/S Ϩ/G>  `E VOM'|.9&b`buc!0Ⴡiw!N]m4oYb%+ \)%Q|i S] }τ`F(&299\QH"bj3-3iTL0 c kH/L o*(.4R9]!N!Us_ 9cV +00!k+#6b29#(`idZ|d)%{ItۅKܪ9h܈,ů@ådz=nbヨL^ۿpy[T!zHuC;gH$t½4= IuCDU6 ƭe9S~(g/&Q6v!!R2)9r(Wy"YĎsR:aA}:6݂Xz@iyDʸO`'b&Oz\Ї[›Z ?뭖5f`͉ʵWV,o p/g۷o`"oͧ`.Osw;lQ{O]T 6+M$?[/TIg+vdOqCBbI[tNV*XY1$FI/g{$)@r-Up^T0$nfBw0m$C#CcM\@RcEݬ蹸j9pH@Ȱcұ@#7$-yQj{k!?Jx ?{QuwchTwSˋ.7_&_f&yE"\"3M^(2J91#6Კ|ҝ,鴼FBZkbX zeA%m^P6dMԑv]gqJӺca!xҡ(gjyNJ"& 1 ' O΢\<9_o׼h ]xOm./)SdN@? {[gwh St'C2Ƅ5=vN]OvNO{;%\!SCdks&$RvJz;3<"Q;ضOJga4wc-<W-\ \0|ܲQM&ʍ<D {;Q_wR_ ɦp넿rw:|Ĥl_}hX21Ĭݤ?g`aVa+a<& nd`^ݐwET"G Z@ Va'V<8,hp ^\1wθHCsh"'p9-y~.dTO/]ƇJ|x@fQ&0o=hXe;ƭ.dN` Ms'1r r=S'-Xu 39+zJx85gNF6B0ά4GA۞",{. MGG4XZp(VvA_{k ⟄oU|㷍J46rz (1uAYY.3]EnͲOܦسCPj2x{n{!ᒬ}ɦEJCt_ ^VV8_cKwc˞EDd4a^yy}0y^[oEy&:-QnzήZn.Qy WW0;#|CI,! eT6Y!,ɢܵlaQ-QZ6W[a2 maW-_4+P=tov]]|MJmEJx!׊`\9 QdW:0%[8d thMٳ:-E=i/{n& e,f`< uADL椈QQ?[lh`(UKp;HR^>#gWtŎP1x_Q{?t~1{cqūrxO&*JLJli*@lX鏿}pCrVw_ʽX#/6hG9[rk{ߍW_g& B'3]01HT?*JK}% }:V]prR霓/"J v`͆[ @ 4'E{-95ftD:ZFN:#HYoaq@U8V @zpF8ϰ7 \ rY,#1C/qJ͝pȼ]zV vFZu`;Bu8&0c_ _kIkԔҙW^p._lfڸ+&S8L?OZ*jkNJ=HR * 5q1:c_q>?0 |}vsdf0Usz}&#\ZdL?Lk=xg(85][ v*kH.f$FJx-߂N؏-Iwq3v6(oЌAL @7_X´4Aa+eaCX,R4DHiE,8`W-_LF-W c| -zWOasѤ|yNz|x+X!F=zKG)M·Dy͇p=ӴBipk/=h\e_fW<ѭ쬆fi% 5uqS3]a eh0B~x$>1r[iӕ' bkNܑc9\Ӄn擋iPI7>$Tv\o.vэ[>18זCĭV:u Ьەl+j;ᨨN3[.>68sۡl#qkOJ'%W bz>j\G'_Bˡ^=!̍_ ͌Z|><﨓Y!+nn<:80R1[ev Ž q.c  {6 dWL8UtX|,4 *X4%&#[HaD@{t\I#hC C9HInr,#8vԛ, ]/}5YFO֒Kg39ED9}@jXcx<\xss95d*ˠ VA[F"/,{brOCa3H%9exY%!ȚrR QZQ%BV:uD(K=▇+V m @MڹW Ul U6G17dXJ|$,7ˎXFV]I?[ilg̶`@?[^;7UenRNαע٢젖'RAЊ`R:a?w jxIl}Ө޼5yΐn~XwBT`EbJ SQt :>h:of0̻{?Lw:O.?|x:['RqVPXO'g_g LJ__hAtg!K{øO'կD?+EB-J-dӑIEC?m䔒 [1cݾw¯Zwťn&wqpӥ.[P.I?=83Q*G%č]>HCm_rP&vkY}in;=Bv[Ϝ}c9%;;LPnÚƲ`gXI͚]Sv KeS/y~ճԙbJݱg 5r!9dʣW㇝V&*vCq}87TIɌcT"[=i%e'= } fTaJf/=r@'fcB>}iNs4 :L Q9H;5wZ2B1a5@.@FO 00<%!Sin4rtFb L$%eg&ٱ,l\AX%C@{.jb< L"ntހf0+DЈ ȗ `hAr;;۾<_Š[txЏč@ࢋaߘa÷^Hc?yKtWϡy]q>MW VTշُ|*88tF`y>gEe-`p{n-DHr74P"K<\Ab}k!HFE+%[l'R $;.R田lov߻o/ךkC* dJrTTH*)b1V(q) Ǽ")9"z%xZxP j(cmom  g!7$ uJNs~@uwNL+*6xKq, 4v1ehT61-_sgDtIaV93v)ӋisX|(,Y/ݓ q(C+\2UI TR0^J_]qծS޿ʸ,pca-`02'ŅTVc}m ]JrM?Ky T3 B@'*-(Sil!Ԕwʀ/)|J!x]}~r:SGt:`SHa c| A_g ,o/Xv'c׆Tgj{ȑ_eo Iv0f0/d;8rF,_QbwLb"OUb-ݺ󑗟>dqh"{UBwUK28VS!G[8-Ϝa)w.q@by+.>4Ӕ@q)AZɾS$k5Y MmcuRahTJgѼR\uQܙhB.Dw5\v| 4{EFӈ9zYp@enFL1'P!dzУ A e\@-V?n8㨿v0Vt~"N"tNG \|b.-~[^9 cƣF<)M{xh|]#'oe' m>[SIXfD9)@b,٢Ĺ_]bwzQХ%RքU)RŶ=Z?Iat4ȡ[fԺ=J.~DGc =% c's$ T;^Hb3BɊ 4b0zOuQRʈrU+iqTUM$52PJF=CV FYb#ٻ?/Fr!.X5cM}KOb7\퐋hNNh>:=摢|r油(оi%yC]y-jsON]"WqR}BGq&7-i3h|tɭ6tV;:O;Uv~iQƗsiԂrmS22swg$ *)E,G S(a߅f=/a꾿U~|oRx745ZQK0Wb|$vB,$'l:~ok)Ef0zQ[(iqd@ _ثD3xo`1J ;;/淳.h^,iz5 EO>pu{q9agK{u-s@@|ʶ*YBs|kFZyQ-Zt6Zx^ն,m³'zl s4Χ.\zpn'9>)GZO6 lepx z6]ev<>:]v_&FwDIhb8h@Wp(dsaLA^`+8%*uNbl¸`oI f^g/6㿞LJcJmeO^Dz/ψ0 D ̨դ%)R`$`+4?9a!S&r#k | ?\t3yZLNԊ$/}q1O{]Ry+hW3D1Q))c\(<l d(Ks_B;s*-_:D_Ɨ%i|IpZI.R]P2Ψ}M[ALz7VK!+E,ѹeIt F%Kx v/N%i-3((:% dRR [ёYhYZRR;L;Rmٙ$E9R5ZH΅!?W؂@MUӍ%_Zejci[G|?㠴*c4N7K&c0`R0xwN|I/1vC^jqI7:q(!@vB{AO%} ZK!eT;#ZR0p: ].u$H$ zcE07Ė6C/}ֻ]|}5 /JDP5EaʧPW9-M*^<1 :%TkA ]mQR4 m (JBfBG*X!HFJmHgL:Vʂ (mFp2(\h#6 CHˠG稴V`ϒ8Cheե[cqj6m.n)m$*5;OYy* uΡAK$j|QW)7>"90b醒OkAs6YØ:؞*@gr DZ+y[J$2,n !ǸEe-bGYp+)'LXԷ$Jn \u$ZdR'+EP0 Z-70SiY#1]UL&V^g7׽bK7j n]|_|¾* Bz=O&Ŋ?^]].^EXr+{*H og3l FRiff5"'_Ș$)}.?O(PY;R۰zsJƥXa;tTSA$aN ͼCl5MN,=;$vZو("N-O3𕴄*@ mtQAQZAJrA R֍ zyp%sG Z1މ$%8<4n$I+#$!tF5jQ,"dY .3چ_ e4 .%PP/ud}{%'+Rh? JG+B3SSe0B+u6BPPU x"8!ҎLK-=3:^;Kg,X*+n%jDYhxsCIJ:Y hq/l>ڀѺIKIG8~Fn=|$Y]''$E)~ lb"̦Ad 2b3 soH`V3MP4GwZս[AO%FWmv/ >i]m]Ըv *Dۊ;hY$U66:86}{;]Mmp{|K߃S|%cK vFot>nI0"AEZ󍃊s9ƌ$OF4w# _puU)v,DǯVU` '9xzpfn'`b+j3<84[p(`$ZrB6-O;~j萑!hNA$ qpNFJ h=zSRLD~蔯׀Nk Q;+Elk9\ZpyBC$ cY oK61RI{)QR(+6Qv,4r8e-faZ ԀءSISZXQrq;] qV཭Tev2QmĀo=ٳ9 ?dzNːu;HjOH ޱ^!4J+8#Ah y};{[ 1K"PVR 3^JT%#U ZT;ks,93 c2rtU9&GJ:qj8J 2"xYT91Xc-d>2g%GJNݾ F{LI2$Qi~O2t6_k *ٳ˘r̩\mh&ʂ ՛1QMP#eÁ8(-fRih Gw&Ș,Y"N:JKe.s@{?]|%Qo<:޵6nd"e1;K0,0Ad2nز#`d((i)c[.~_թs\ u緛F/oF0 !RO| 0KDG^j ~;xc4,_Q\$HRBQAs5@Q[|ZHUJ(*4@QŻVzЄl?BtB&&XXQUzGzX+*V5: %XtFNh BJx1 z kq-B7)pҗ5+^u,*wE(g%6,8ߪk~[u|-1Xݎ3ɂn"j) o9a]_CF@Qja9*ӇG+r\(KK&k4pz37a2i^^yyeyMFw YN(fX +Vֺs>ylH>5ͧ9ߣ|?8Gc*4@a:`@$`A? ̮CᡟG~Y*.xʎb D`U`r$aM9C )fqn\zx-`OpTƠyYDzg9DL g+*( 9_0iW@kb~5[T$SXg&!dRʤ6b9R1("$͂V^{-tX8jF$Lt{7Lcr6@4LAĽsX=H@|O%$?<Ĝh1dz{ϣwA"5Ulg ~(H(ps wPDix?k!\kBaӋ:כ[%BP1^M!xOuyc)Uh8blu8 `M1ЖdEU6<"!bԚyȕx?+' OGVe"yf:`8oxqRme}ʀ.FzDσ# C,S7dJ(MM-*D-xBʗ xY ,86+xr zOT|Gc&f.Y ` ߍqe4{;Txr#s&mx=USԂP>Bk(D0~RҚAar b(FxseG`{E) bI<7\Q״=N69j~L``fz8F>T0T5bÚtZJoV^! (`֘E^!,\kFT$eb:ϴ3{ ~D=\bc:H)`W5139U$7J9?Yͦ3v/W߾̚$vv/ !j \SeC-<\)ª\DW`a\q"pb!)<6|qìd`ZcJ {\ b,-SN{ ؆_;P~<5?2;}#ĹsJ2Ev˾Eӷ o[P ;S!E<s?'w0c){⻋(_R,&'bl͹f.l$K1~t.pJ*fu2YeI& OamhK|5Uy8 9ɦaNV=5;0%O *JmX%лe '>x4_*7[д&J#Y[祳#IV8isȋI%t[+l]C߱, Ud!3;vĥ- Ed`v31q s+2AJ}\cIJz3 @5ȟ]KӺtNcpւrn2 &4AG. L>}4;Sfax,-%ԉ6 >U)ʅ~y(607:,2Zg Or1JrDwG].F.K>zp"Kd8{a)ܭRJB5dglYw-I$|d?Cw&86npYcXb"Y¹ID ahͩ&Li(235 HalYPq-F|L5Tw~:/1^,ՉW "S `Ȋ@m!WN46{YηENA~kNl皮 b5 icIAզ<U1*\զnېtkzjP`:%G1^ޏk2j=;,nyzԂ\nþA !QB*-^gu-ԡڗm&#BD7A !WWBRcXIIX")3<YdUB,, .C^t:M՛C:?]ㆋO/#i;2rԸB:9tsA7<m +P1{ IZqt4G Z +6.c^ [m *Za!6SsywmϲNdt?֑Q(ՊwlYC:hZWy'ͳ[L#'m~˯G=2tse`m(%5;`e>X'h͵eE!c:35٫a J"As.%:ﻻ`׈ Ny-[4g;nLkFObԥNu~]}>^ƴ˘VzJ/i>ux$^˵4QO^q$DNdr͜Zq/Z}}ӘJN7?.eBn#@)IfSe6l->7[qV c:hZKy@$JR zE f=*S3}TIl- P'k>l$+cIJ#y ,ai(a`_ /kGR儆rKDC^MF@6TN2g*׮֌jӍr/X׳ M9K(mJ[-܋s~0~*IiإD(Z QKdD˜o))""z[WAlw7q>t7W㡵dt+!A-< Pe3)]f8[c6n񦧄ӴkGXx}#ʍ<3jho#To%=EgmxFU#Mh{;kVf"UxPuf@^;ҌgfR.5G=Y(2;҈4gy3eKn:r@5'_ H e2#<\`,TR^\-*!eo8rɨbkρT60*2veFQñ'D(ɏru nX:%a[Ɍ]$sZ3SmJY;%pYTsv>ڂBZF7w$CuDœkQEUV(W_($iYf)|W0"f.Ok.q}7efv].a0>s1mh!`q ThA5]`۠z5 q;1V hp[=kX :7P28ʬ 292xgz7_rNp™~8 8w$%{%;Y[-6E`[Y4 TSsNE^#ni-2 #߰n* |#B_5(DBu,،I9.`u2mpaڦm]fNsƓ/iLwrM:y0_EZpX?;jb_3Bט$1|5!~<`o蒢$}??|*?72Ĥב]j.> gH!]yKn[*%)f\ g?1/aIDCV3v(ȤDJz^pN!?3[l]bkeݢ,gQFp酫92R ޔV#rB3كRk"!Tq^R{ȱ}){*%)%Ml{DsNvJSul!YS;bB$*6aDc%Hb)x!{<JKRufS0Im~ez+F=R'SA9MHbLC0h%gr:w0_1%H(܊1OjǼ8ZFPpeomԂ.*~-yQO:h-BϛbtԚgnDkJ 2]SZnC]V5S]YS=l82J>o/=S"TuN;3ХX49VLQ~ z(бqEƔuֺDY%x,\ֽg)}|W`;4?]Y?9xJKؘc,Y&5VQ{֕Z #KN(9EHRm fm/z) 80CE=zj %^akiu!3QMSfVKZ qAO-U4=KEO7<tlۆ` X%ǡAλ(^5a1#Ci [G<9y~Tp;[ )lH`G (d <D,TI;͜ǖkxɻ9FVƝ^c\z Km8u"4*#h2 dT:p.p]դCӳKN_>ɪ\aZq -NJ+Lz/5!J.KJrSS2ggຈJ)zVTL%愗C,Qې)qMiBQb@+W_coZ#*y1ш9`1~̴2RaE67Yw@xJ \͘|BuuyןOtӚg:du=[b4֚gΒTVFnE5-܉/19Fum:?.%Q#JS_qYyN ء0jL4 5m TS`oF;Ӫ O͌a“_Ǟq>M7}~xN}b':-BO)eכqH9Os+?pZ\ļH@"s sFwbCZJElO09HjXڦ=Kx.d_fANzqt-/@;0R%洔&bHq6iFj͟q|>fF2 C*K78.e)编M*Nj{E5er ā& ]l=<}z 9T&k0M `ONtz(K&=>3t[+p+Es GΌ)?7?~0y d*yđӫ_ ˪L 36r&sDž Ӫ k\hp+ϴζr kR/KƲFSj*Aj+'#*'GO Z,k CVn909QgoZE PԱ5Qd[rQ)<,2(=xRgvQXrDRey@ UXLz KKϒ~%qa$ ٔSښ(%8 l&BU=ǣwReM{`BTd#\P)ѸVM'\WuRlB#$'Nr`TW,l҈fu1O%0@F<Љ NcV[V"*v5*w`mU B]*!U;!4)9p?866qB09|דɁ>ާ䆞WZsgN0ɫnTwuwW/qH&+t}-n(_t'e\ns K8=^祧S@;AMvQЄ)&l2۞~͐8_JWM6q3ĪN.?T EjVi^^w ,Ox4g'גԘG Jdn XzӎBwὸr;y!PLZf:G '`O 'c}ڻf,أΧtlo:Ɲ._|g3ѷoCoF=L}$TMRS] &]]$G3_V;N)T[IVqoc.c:_sQc1MEoxg) \op݁8h)·q:~u@u5{:~4'o|zN~bLR%Tt\*+g=7rl+Fx0gyҌ+m2K-uai6(9k0DŽ1l@4S$XbLO i%ST-ETH>|of+Oa<{|7lR1-qY>yLY\98͓?]laͬZ'lILwWNjYi%a%a%a%E.\ݓ}&H'avYu9e(dىCϸXFэyWjD,!8 s'x$@gu_n L*y&QY3kyJ 2c) If4VdD{?G Uy3 L'lLhqA [^^z XDzhej}Tr!8vGyv1w9%ACjeO8JG4Q<6B:ߵgFanY8rV9)J4u{J碙#aksla;qR@C>LK&CZ*)g뉣,aI :[r+m*eds=  #˂OQYlj)wDFhzR!sA1g9<[Ÿ˿t ]ӢP]LOk$uir$(<ɐVz7<^Uo"78 uYōZ@T + y{w}@mb$di mR 87txqi:cz;}{oL!2L4yq0Xco:o{kV3=[zW/hps-Lfy&[[;<ʐ7Hd'Z[Nyi?+;YUxw5Q&W엫~vގBt gWwTwMÅȇ=}L( >Kryy7whr~/'DDH/u~n' Gmop}ޅOq2Al't4~ >6P 1N_Z7s)p./0.SbDpi:lzgiywMv1imi%E4TLf )ͅ\Z+o'H6BK6w`;{Gu IP7dwv<@;{7>0AbG~wer2{a ~-x1U!gϘR^,Wso4nL99+Lٕ gVU4EE0鯧pY񳔩dωg~ch*~Hȉh!9M4B Ft>cv8;4&Vv+ڭ 9q_B5taLS;Ӕ_}onB?dۡw7ܯR'W"֬:^S(ҹ6PpYo <o"2 M/_˴ǵ3k܃0I ]8PŽA<؆{>}^>=bKqHCiq3= 8/h?i2!<ƈS\~,2K:h7ZqrpHQTPV)ԯ^_v:O'#\х;1kj*ΓXڲg6!cvr>ط?gr @qz7UOuꗰ&mEe{B7cy9t;hԻy!PŊޠ7.r{yPnsP|7ǟr'oڌU5uZD-&&Et"J0ĪigTSL=7 o9PdRǕ9 = IB._)a/NwnbIC^q'%EK.D&cN=8t[!C(@ y ndhD1D)3 S 6)l@($C Q;4!yAkPԦ=rP>tKxD٩Rl(FCpxGe"%MdSJ &D$OwLҔw("I) --4fթ)=] .N7uV65A=\dLa|ș{/6' Y Fh#yhTɎ[o9&3g=MQ@N -3P<*_Jn6]5 ر!l9e`Lif7~Ё8]nV<_3ZPݞstzY:hhnNw:uqf#DgFYk> AT%ZbQ#^AU)'8)ʑ/VTP~43+ RҠhF@BKih(#Es"yJN%fIHT ہ8`RU% P$D!)z>)*|$ 0%9 *- M™.X'iGSѣDJ CDd)6Dh%̩d$*M~g媥QUEk?~bj@[˷CM 794p M #NՐFrGƈz(ё6kP)c2ޫF:;҆"Sh3Z!"ɠhT eKb=$ . 1Quh]S)T SV+V|kbdGh M-wOa&*S&MbaL08T kJ.?ݵ:£ZMIv5ǜSAxos@ڟ`|c|?V;pbR4$s:}y9;bT*ʬm Z4FM%.ˍuW1'1rw7C,Tz>\ )} A/u[|X(;AΐF5YiTvrUXWj}>\DS^q*s;}W͝|^jUѯҫIפ&25A>~\,k֡ 9"FcuIF*jRy9/DRjM,qە+f} K~է>`6^,˸{g -: _oLyXߘ1䊎xv^qwo<797 ~G*v}iUϛ>v1졀Lb  _>289L?Å VzM2; 5щ K{[e-аӇz2\ FP#mҠ%4x"I:9C-Z0'~㔐 Q<7qqss|QyFOMk2pL8ZtKK }%}x7eRNj ,.t`]-ǃa :JFI*x)V#w _OZOJ&X2F'={.կ\b7aYAVmc:0,%pP+V{Ucw$a1c 1񽮚L ٖ{t3ʗ@amuSgZ ֯+~q(mJjCw=չb!' -eK%\i$76OcΥȳgKZpޚ3?1:]pS;S-D?'IWQVHBE]g~  udrN؈[J$3&tIka^G]ADخpAII?dSH9#VMQgS,:VbP:Ny<2zx_i$O.U2Պd{r[-%S>mBSM'n -nCHȟ\DkdJ:vӭʌcjJ Dt<ݖDC`:Vb6E`RX \We ]JH5 HlņHX ,SD 󂫄=~I EA)`m1R5=% :by{w{\,GŜ߼+Lj}8lD몶Frz^~uA;OPk@O*dsĒ~؍X9>v^Hv\$hRzC.b͠s` nQ^Wr7 $r4eQjZT]y& TzՂ׉!oGURz6|4Gk^zx=cՁ:wc03V G9j*a1JEaܛV zn&W-mΓm}*A (&ԃp) .H^diyJZ'=n\K"%@ R(F ,p삌]+éVn~p?TЩMN$Tt b/'#Djj\\THP$?-Uu}}Gmd^\.+$׼@ku^gΪ)FX,.?\ڸOI6u2M7kX Z`ct`D(8.P {ld@WZ0YE19j/;Gv&ǽ**&!5a㴗*HrL  +LY-z֔ja^Dݺ!^q< [qCݺU;>эi9"N#Ćx!F_\Rs%a;@ݺ|<}ɦZ3fAk;//o90w{?rS ]e$IZL[FAXh4FzPVD9{l%|k;NRV5$9qe5 39#H,Ǝ*e"FLǒEYpaUF:w_) [S;#Wm?u*иYh Eh<7C6h9<m*їP,?·Ro>b/O;'XMFX32^azu%-T'ch$$%Ec1͜ytrJG ħooL`L0PS THN[s"5aETsM^$/ ywWRk>\ghhp3/$߰K'*H0 C ^k>Z#oM+C쵽;O 2f\j 0ްJ(xlz:H/yN[pƉ$L2V + =bO+2k :5VH2Y^-l,*O8^m_x]k:qiY=H7^ޖS|Sx|tWQԌ C36ƕ3(fQ#ȁ<VHĤd0H͈WĈ0c7_t\8 7@u*+- P ϸy/c3B]StJw|djEcE)Q`'6?x ֜5X$?{WܶD^AUAGRIrR*2H e+@\ĪI`3=g~)&]۩uF$ 4^Аz"Aݓ9IksRo9 EUt/iB/)jݷ1sdmkul2 =܀-]٧VK4{6IAmTW5qvk[`yj+uvCNhM9. %BWUwiڈu!1\LVPzIk*PdSv&R\/dK ƃ~NJzBȸǢ9,E:d }}<7bBsa!P?b|Ͱm#ô6(M40,dze>X}lVJO~0~ Zƽ΂{_3JB إ}R2c5,BM|?AJ%iҐbX\{EM\ ]ڳn4ƜqZZ.i tS\J9k&MmL#%MxA ^Y Vͷg-3(|;|[nGbof^!ւl^WG^o8j]wG^2[I&o;ޯvwInRk^&W-!fbAʖPpȖ\˫NST(|1eV 1rѥ zqiFiEwnUв,a6KSʼFv45mVx.KZqOGԿz}/Y"P[ I$\pܘԺVT*8ljtJWuJEC*宿z==> Lv7Cѿ]!(P=yyFYl ز?ch*Ms(:~è}jZAZ?|Jˆ0D63nTI9K+}'`1ʨ~C77e'Ui,0bS7 :YSGIuL'K䊘h! ƅt- Qʽ_6Kb^;dN+Ks87/f}HwHNYѮufD,emUxt0٠N\ ~;'VN7d},(+u=z]m)(͢?v赙vExaQ4LM2Yو<ܤdF{G;DҠh$p=sK=^+atէn$V}/q-(_".aWKP}`! )hs?|2eXio]d4 oޒ)FwW+u\I&Y&G/Hlޅ1.ŷaxx4hA;5N؂љ_?q>Vv֭q7ޢ'DNrm&t{uR}SW$_AoktOFϬ5TѨ;|}j8 Gx20GBD2`N?QF?1kd?ź%ņ]ŒЏTdp\'( GcN4$1_iA#w,SXOƇ1'E4 ,c;v4N0mc+IğGo3M'<9C{~ݏo?tza6gGZm&6 \vRVfLOk39:^C&Vro Hƅ#tד^{WxZw?\zs~{O~Ow޸zwq_N]\f{׋/vzwDX|uyyi,ۻWϮtzkxOw?2A~w9&G'O7_N+ѷCg8yP]y<~MrҟګjuNʵ:94a`Ǥ/nj7`rQџPw&)sJƛd!+F5|vMC $Tf*SIn}:~Zu $+'=vflglߝ| _~~w#:<w{ 痝@'JZ7FɓR+߻'1AH=򊞢neZy*a415MNߍ[D@1i9L]8|SO0tAk{6=0p,j&n7f*My4Hw$Ԅ~m Df\Ҁ4HG{H2Z6$P=rYzNԜ>q'"tb g! ۓ|k|9OL5ъz9lE;|?E T`)hVϾt9|>?2Ǵ?p@E}4AsUd]?|PT|8&73t^dza׹r\.J܃tvx8x wÀ61)Jz@Q{O%cʮ>m;@hC6w<~gx׿<~<~3@F9)Ϩt|1ñjPH Xcax2x< y|x|xC<^P<x7n3ƴK/w蓇ezt3} 2/L 'M!oFu|z'?29cqP[eXs [ڮ@>眣Ǜ 1jRD iw8AaF9B0 sxspI, ˟}0zƐqڞs&ћsO4u :9]i505LdN^ءӁdAg,iǣq 3zmW<5܇4ꇫ;eZ]e4& 6Jfm8ԫG^)עS|CR)X ,8vY.GdZ_ zt5z e@NNdo>ꙏ3኿~L iq Jo8uE r_iFƣ6i].Ցg/@ŏ1HL<3/oK1v)u`" <htp'H0AF&FgCRZ&&S`(& 1'v= ـUѻDuMlєכ2bnC9s6pFsس1K1:H" KϠ9ʼ`83ȍ#mISӥA`XJS \ ;"J Py>=)aXQr-0i;~DZ,Duz9sTŊ2-,UyCr-&sfǙ3QO_wh}|lY* c9기S0mϼYbd`p'6TI+鹽2{n=&a5ltymDySvy32b` {E\+kZ3V:ZvݪN;0&yHkuuH7OcVMnBw4AoJZ cʮdc\Gw-ݔ|$p5aRwc1SϧMįZ|RZ} 5]U7jpuɶRFcAi]J+*A͊CZ ` 22V &"R01SS<҇jЇRAQTuTNsHj6lleSucN9IFor5[LZadE">ԕX')*J&'%KT?XZu%TR+R[mf[ipi#j}:l:LEˇ!:utmTXZufRZ&:nD[IIj!QLh`u}A,HY<'ZM:V})"tfѶuM.P׀%܌X^$JUكorހ[ eD&ŋkDoI*$a#V9jWffГ%m Y[aBDY)iRprb5MNo2\ʥ6W~=£Y|dc*U2c*e[B6xEzC-yEu 5YpKd|ۮЙLGI$ ^֊Ii3  $\~vlbJ9GMGfF\dduxQ*:\S aÉDM@$uYhO>!~+$2ź*UPZR LVmyX~2MoP(u EFG3iHzRZ{s/ANޕmlBCfڳ/!vu vZM C7"JN =CJ2HEJZA"n3gsldz+$_|tծ6uuq3tѠ?6 Oo@V!c tX8!7XpTDY-Nd%G)Ae QbFV 6PqAoheT`M?\B~En30A}p֢4i*ۉ=gs)sAA# e߯FԨ9QY, دE#͕r>%͌IEaV$1D qb&.Ҙ;Fc*XK&%rNՒfEJ~:y!O)k์TMB+{V=sh ̎Q6m` [)MI#DQ"18VIl\nA%NM%fi vwQ>I-g1O$O"fDp}Oj+{JjԞxhIAL,rVca̅RFIK#&?D@g/Ii*MJDW'QFgOy-@[p㏟Vo!D|;BKeYhONO.NZA6;lAoib >N"% Heߞ|닐ȱ.-cR^ S0HsHL%(n_Di^&$b*%o&еo&Ij}t2jʵ .MC'CyHQ'K 8 t_ccwm䜓Ֆ?)VItޖ^~`|9>ћw<,K/=0_GXu^[/Q !kb? (3-ט{zЪ&?6܄,_JUt\#7I#e~a\cAۼjKnʼ=‹[VNRvhCj;Z^OS)/m*hyLּCf)+wbpnuv:E Nfx<Ѡ "?#%uZDpy:dsLAuqvI Jʁڻ7ǯ^.eWm(,'8Oc6E#-}u5c 5'kBɟ~u4#A>y5.,;ԮObx F12dFFCI9%:fG.Pmԕm!RXmU+OܪĤ'nm?\ tY-Z$+/!8–xsgˏgɓ8dTFiBokLPM29kg+ugԛt ug͖Da,xCQ,q˭e<elOu)EZwh5f uYfNA!w SkOVL.:e2d/?k^'ɅZWL)}AR)e%JA RBE6 udi 9f@J YeIRգFd=9`D|sZuJzᶳn}rJ<)-vfȧ65Qf|.ܰ՚ʪإxWaZ;v*kDք.yx"7juݯx='-.i(MWG|!_:E/II4Fm8&f nQI- 'U`$O `J/it\4%߱*([ ~cO2&eƜ"좷J˦qO #`NW^%~:}.4+ea.Gp]l]lH anDmu~Ho 2`N,_4-oG]ga0 ߾poׄVs2k[ftMi~fP uj&Wni㫔DM72eRB#W o2Qڏg$Ƒ⎄8Iȴad%I1 zPL%-QWRFUT5>)wYg#O?^u3RGAԱ݄'fMv#+rDm9Oj|h<[rojnnf3ܪ4@Z͙_"[IGE[xaˬr,v ]`*T:mdiΧJ M|gWnY㫔LEv/(n]ľ'$d-sL9IKylv\RҼ,!R68ܾ2'o&랹tŮiĀg7FC FbGXx!]f7@7#ܗ"U417yiV1b`??dvsh~FF<}V r3Ӟ+s]_7 A_>xz^n&Դ8}|ޫe#$Pݮdn)8?/~>Q{<7~;_]W3:'o7{ĵ/޾;)~ݽ{:ˋ~~|tΫ^~o{o]_l֞z>Mґ'e_O^:}vf|?ꚴ;)L@n ˯yk؆=c067hޘU;(w>WFRӓ/C[FYzex< er1 wa.ϳk!$y1aA=7tє3f9 )2E?ۏ_oNr ɦ9}w[[ۉf}~/0̍u70mڷn*Ŏ9v+a2,>KPs;L ~@/cGg"$ápP8A\DPj{m.E [p WwER6M|7̍f#2:' pWSwt [9kC 胒Bpr1fΐRbJ^9$Βă$ pX]߿o*6E4#~!9+WAEB > qsq3tm17p~~/TK'O/gEFLprO[QVCbiY1i|6̣n[Hr*P!GI G[{-/x<=/c36$R1(&:6uq;&y8I촖`|#q k4D Acpۻ )Ȁ愆$QZ&L@Aḵ11kX >DI< Bn8׊X0esstJĝMVd;\˾,ǩ6E~X$RǜēbG;?=}|B0[Oyd"yp4r瓞! wG-Ol<7Ejoľ;"Y8G+2V5ǴfWJch) w}%* SlN}(7rSdd ;wVG)?fHNSqcխMhMp &!X\*’Pa'JH޼61b%6482uHswi1XJ"`U?#(nD` g`PӲF[- J AKdtFG$XDIw3'Z1qb$p% ƅ0J~Dؐ&6v*0 C4a0"3|$8WQ`i:šT^$##U#KN3 #{8{>іGf j!'&!<{Y4iЧAC i*Mց3J 1We!nu#H (]fc; aa%Wsr)p {u|T 8o E_0↦'{uӴ25:;]V@G~ N|[,%5?ϧmӰ_:hySٵC h'Q75I\r˫m\AAďB5<6bQrT&.\PU'@,:RU'uL_%TY%\bǑNUsWVò#sϕ-FuKu}A\8` O.ɝݝ!gwB3_UWW<FNѿ}׈9%ٕd?= ج{_ :j~)`ʌs$,3I&=gL+pݦm 9V/_򼴹T*Qk08e i#QG@h> S)I*`)rK`$UiHꕗ]9#0di2x:m$nWa&<[DU ȹ8Tbj+LcV3wzۨ g5i&Ǭ>:ʄ!OI4i:anWD f@0JU5 ɑ+g]+,Ykj{ 7kWrUęJ3xn*-f?U8$KG'QDJRWs6x2H@j6^(r)ڀFIk.8ɢm_fusW *dNu6FV2N!yR<+d9+L E龰 R&%:t4ţ J]I Z/͂{B (غτ Ly2D~҃JUgN>|4p}΋%CNJ:mID$|V'Jl,jqIh`MMѸba8|Y0zNUwz[dDey#HCpfZ n`ᖃ\Z1}[ rЛJ{9./(VìZ;?)k4O.#m#ONd'fbHߜ7|OOTк]u"qtG aq>-8#V3k(^/.Z/t.k'pH(wvYpzK/59Xr\@Qf3:em,'YfR<8C̪Xӽm[ ̺ApT'n~>z!T#LphH!&22ƥT1zsNA'@*cภ8a :gg%G*n";,eH2L>MǜĪinfUnQz#Wכ3f2WUsЌLS.iHΐ+ 6rh&۹Hn@@ T1,LROwB%F&([KdOZ#ȍ1GDV0*p$OFBP0ns*hw72J>c@P8- φ9Ґ?Bo1XgMSCx !>yOiײl 3Pdܶ7efaШH*R/ʻ1+HBXRB!{ڂQYhہJڵwm7`H΅=\H_o5"gSgV=Vg%Y=>Yg},?)>vUj:h[~دu5гgvhr}w*lQ퇫n?/u?\׷36t7o?xƬ!7b4@S75D)xvM'w nY| v? ÏQQX‹zcLW3o=ꥑJI&)gV*&Ά"OB%GJ ߼wjQZ^.=?|$XxA/..˗dd Tjqt؇ЮL{z02 #}18DŽvbbyØ_I+j*cO2-zL K}M^$<**SVBRS*Q U\=xNcKHe//7{Ov@hx>ݝ7zH^O`P|b=0'wؽ_@n={t1}䴿)< -Ï?Knn._i5rD&glZSft6K;Ĥ 3+اQ " AfؤP#̸6dd~ wKCH? rFШp|M6%w*9\)dWN/ <8cn:q,B>_X׋/OB g _dIKb°)g ^ |x0,9OfR3wjճgU5˱F{}+-hۀ/΅&6h&v_%5 s ې/&ߋcOBz ̹h\/5<}ajoβ~!w_z/(3׾נZyWWy>X.P%ۜ! ikf&IQZE:d+J2 %ƥs ESzM{C1`F"ϒH%`ķH!rQeuUJhQӿWNsJ `Ntc6`1?J_I+#Hx*j%j&Qj`ZLl'sR{b^>vScV ѺAހ"r( W$9I^r -x_P\0;O :v7 mnQDњu:805tONHǜ6Z(҅}A"sJ .8':\ȱZb(1`N`)#t}{n~T~^l$V~)#2X1,ITHIpx$G4YdV{tgLv6tKMݻ}j5,7o>=Ҍ:M:z_Iy#ShqW`w [˛7AvH;+|;oՔ=LC6&Ô>8yY'xKhU7d.77/ fzJ+-gP[ʌ֋1Ա$<%sX=y@› kVق zM籔ʃF7R+uC5J;[X뫏+_RĖr_}ɦd/iCߛ?B qn u׀p`|ȼ|gK3[ߺ@Zm=:[[cZ4ǖVefCi#ۿs13Uh% X-7 #FTȟ곺_X#ڟ]$cI"Mh\0Coڕmopi&` ıE%q^?k;h]oGW}]"v`7#S_#q#Q Iq_5I#r(pZ2D35WwU1|>M.IV'&݂L4e A+wOBoR.K#ε`Ӽ0/j%2]j)#e \䲎)Dj~Q]~=ɿj-ڮM]@kDЈfof" DVĮ֋BfOm<6L }7\&s܇^W6[{eZ+gҞu鋍Nh]4 =Rώg|'W,E]S}uTۖpLtu9]A\A ›4%Z~<12ysكKAkn6PJWY@YTCL^hw9׹G7a=-t1%4-OF+X3*,+԰`p}]ڀVZuWè>޴נ*oWȊ'a3]+vd ô|z>ϳkÿ "K84MaN0GQS':,=C]zގ[8#EI0 NɡV Q8:H F8Z¥M`x˦H{8gqqNKs |nO"mf2h?ǻo|Q͗s-[Ưm f_Au hO͖ 0=nTܿͿϾ}{v_._=BvqE}S:EBgrj)Y@q,eXx<cqc=E&tnS(38!>M/TkOyTb `%"rDdRLrarfa2xsYGFH5j,Z 0ts>3) DMNk/C@1+9, &Nbs`I!hO%BM(R.t3 :^bIN JJ$ [[Ji%g [MQ= P |jD19ck!HJFZ&0`l@dc<6%-MD $M 6>ckP-I<1wgEq@/EVEsU?~͜$ߑwo??+\0+Q>+@;d:[*soS-~Lg_/Ö.~<#eťbT7o#lR%5Al7]XrSœ(H+)I&.ûXd*hA_]w[T+V׷?Z3fh1~y{%8v pOdn'1Sng_G,= _'\.6v)/>cAy|xs_GYo´S.Xe}\<˶Gђ^ &$V'˃s8w1g ҳ^Y /! y"ZK$o8ִڭ.)SUHjhvkBB^nM) Ѻ@e?5&vDLî*/<"(ƴp1aYS2`-QRf`U A3Bxo22;FD . X e&4`lj%9 ؞DYr{6-{$, (4R ɹCLH̿"Xj9"%7e`[-$*z/Q'o\.[MF {s}ajTh֘Z_)}^r'R·TkB+wWhRޜ}x, v2vAHlg혎Jk)t~LS-^kpn k_s$VҦ"XZnUk`  2EX)QR&0"\IMg٢.eG-rs"Oez'5'91 'H%w酷u֏w+w'l3QܣI5~ '@QfPeG ̅Zy{-k=5 { 3f]#$L\gLiyxKcmR¾R qͅ X3. JW H;19#1[ߌIS8^w%G46 7]ӻ@1q]**VZ#,b>'@@ Ź1㌢9,J0>bIwr#5=\if6RD 3V".kV.cEAwH/JЖX b1beCFZrŪ`fܥK9,!X'!jvjRe[eU-'02@oQΤ㐃J=~).i~lXׄp%SA}Ī51H*J$mW}j$䅋h-"8+b":uQE3\t)j&$䅋hQhMxݐ!I ;uj*\NfDX*c81OTf8w8>3dg 㟨 .V#ٜO+<6ʼnٙO~f {UI;CN6F9BS\R)oKiڼܡKWGuۑ)E$hDk0섦 ?&ko П6PDP0L{NȰWwtY",^msAL`q' Nőx/|tKg),1rM88l dv $xtnkUMѮBŜ:LW_ZN7o;$;EQR|Bb[ʂBlՔQiy@ǢtM(,\O,hMV d*ȵSt4%fa ;}u;ew?Olf)_.]Q`VdN+n2`$hW3.%;k6.[ E˒w m<*2ObzIt4(ܶ CJY |iӫ>C^4(.} :n,E9YpޮCd=F \Pw.`pa7y>E>mB>)P?.s27חTɤ=\RZNjm'RDA xcNSwlFwFlMe]9Z'@}5v;ثHΎ-d4K(J`]6W錑ϫ`)aG47 tLjRhԔ/i8`kyooo>lxq%w[ fww_0Z 0`dȺbEfԪЀ$$oͺpQ_C t:{(R `JGgZ{7_ewBbv"de^5%G-I7%e1E^Suﭪ(Dƪ {|Ya |ޚ ,A,tpv I`nEK,.7z`^ ڍ="%$7xhJ^=!htHRvyEZ(㥚r 'o4mQiQ#7X$ϓREb 2SH6l +/_&f]fTV798/*Wm!H΁1UH-DjP4(50- " v4R~oC>l?><\T۸>m/H+]^F1YHTvgO\[ﺄ́yD޶h*-'Z^]!UKG5t;KNH|5R_0_c߅s4$*2B:&" 6,uOWL#Q6jqdA}Rn W.ۄ[swʂ@YXœc𪄁AZ?|!,Y\3g(1᭵M MHL Ěm]7To#f4bt :L)77jaBuם—yXEwL5rw= q~΃fn jMyeb 4f-xJT7 1BݬvD8i_9j4BVfeuk]{?/C=ۧڜ5=e ,TS| d4ne?F-nǕX pX7:~-ޚwɬFnvs/UfWf;t6Yk1JFٗ@Kf?f50y%KsbWqs:84<ȓXԩb䛓|\2O傊LRJ^M^@&=ĥUbc%E#ee0ZJQ&KA˯q M!aD"'8Œ{g1`96I9(aqXs9߸#HCBO_o w >ea0`^[3R=I4> JeZR &y^AVJ]f5 ~ u"L"UypLkgU}*p$Ah^ԧ8G}EXG˼eeB '%.ΔEe ^JRA+8JO/?;k=!Ah iЎՀOz>O&E8kIP)RР ȧPB^V;콶2eDnD^Z,RmX^!31K'sw1-.BBt.fs IMZFi=vF7@Ђt6"T*x M'^n倈&zxoiJp0*xoeJX)F!-if Ifi#j ɋ}9EE`by̠z7{5];QX|Ċe~UycĎ5O8RTyR{+ZbOM0FNe"s㘧ҤI9#Y&cT]͋NZ%ObjxI=1pŊ93NtBJ`5~{ȬY0"\Bq\1"3ⲫhFfj?>C4>4|2v\x5a8 )8 ՟$ShU W1*Z?0O@W5eSɨW) >ؠ@Xr!eo}>?CZx3)<;>ylZ &,au[a4p6?5 ?|a1لW1^.G.qNV6wGV}\%^sĀ4pĉS|гٌZW.kIԺӶ5Ϩ}6z.+Kq\X.i{Ar@Cirxj>-vz09+ҡ[ϊguu_3ǘmˏ\ *kc3+"wZJ ׏v{b|k,ȾL5RKDF=B,&8VHn{s@^+߾1;EmRߪ>FZ;Ƃ+l$a(kҤ4h Г Lgy#5U2%"*h 6Bp*4p $1~.qq+q*O=nkn)5k[ vEQ5 9~ìhzUhzaE$n؛.4ڷ}ߵ Y~[ QMMDl0.`Lazؚbo;5}K4V.e޺dкTk$gj6Q6u g|gB Ru&e{VtV#Lj&IUR%i@"z3x-ǼH IU`d0+;a$pLSY)l J^(+ӈ&P'?+|cw]Yir6 AVh39ޒ%+GNL)U JJvu)ތwEһ 5O~%fV٬򲡥h+ LE,X >S'纻t1BhQWKG2!ÃWIX8X陝/9P܍j{cCa\$:?B9Bd9:st Tu-AGεbMHkR|74u 윽qՋ[#K7=:RĚ%/ I+Hht϶4(^A Sn^qBIՇ/EN!2q!# Q9iwgSܷ?К|O\Y@ڱK1\+|r^4M0B :lPJ! 4^#H)G .zճh4ZSCˆ8RCeQ|M56^"4 ZcRE{MP)&+dRZU m=SO$NbO$N=`Vsb uXسψNmǰ,D.B )0 a'A4\`c4 )"2d0(̥2D,` I/8l50Py[?)aC -ӎĚZ.Ā!ՠ.S/-Uy^L%\N(b?D`e *iJ x6.V8Hxj F.>4)FRzNq7,)A!H FzArʥQH['f֐@X1-{ʭ^z5\5 %Zoi_9 ڤ3 OB:C1|LM|y/vXnH_\/=w[i[u?Ă'Uj[AҦK:iLv&im5٤Z/wq"Ylh*0#ͤOc(X0t[Q(ɦN^d_}U/O%'=30ߗwe3^bh h_1+bКK_zEJ<ȈFQآ]3 {|TjJjk'%A*g<=T4Rhb]0Q'PĹ`ʢ oQDq;{HhU(tW9m%@k`gJcweł1JgHEY󲁰Gr*~YHP2Ȩ".3 ,CHUgS#9\ZQF f_lqG#(&7ODsiYɑJRnMKF.sޑ|N*.\Rqw+a F*kz_YtR yG!W1%CM!Ya=xAřqdI2^,i_fEZ4_G5H)ːC6lTR ܸPݱ1:)'Mk:͞:]x7 )4 餰Ƴף̮[-.K=/{}/e8^yY}E$is,rf8uHLӨS/7n͗?.ZRK]TbTkW~C:4i ^+em[_HVS"gO5G%-X2\%J=> $х;Z9@V"\BSnySGO#FᠬckeZU^'ZW^WByQyil}?Z/]xt\,&8*d~C=y/:˞_CjP{8Ě_Cu,֊[g`1'Њviڼh,E9/$-nGOgm}y}{eZK_# uTQ'՛s*QV[- ֔¬BbQ Mn k8 >E@j*75}oYt4 "4 RuBk7ۍbh9MΞ ^{ǐVShphRV3&uIfgN`h `~cb_T7//]o_jPAꫯFiRh*G0R [-{Ad<=[YC.u * \|).IpUc1Mi)!i-JZ(qD;I赀ОdPJYS)a@H Vʬ*WTEq Rie,ze&Q"R%YRb08u1-ZRx"eJ88T*WI9H)DB3'NTV އ39ZjJ"13QuH-RRa Q*I 춞" Hy) r~k`P iLV )j1{hZ+@C}2̮ ƀ rEݱrՠ|[XE]{*2E3AN0ts(k2.x?q`cS$*6U j:!a], ōL,NV6BiSnUcX0bXϱ &% !@PΉ4X4X g܀#Qbŕ+͑j+)FL9%UОgKefoJTb8w`@Df[~aiL&{ӔSŅ4 WA`YN%b@C*/6ՓQD2JֽPQM|AQ!"%F_do`JZp~G^#ͩ jHh`\`ؙR袺TgPڝ8T}UJc: %V-IE*ŠqթNrB-g&[ȅv}NW hU*MLa7)PU62΢9-2! o6hС7Kʢ[QYKKSªUΩ ՒyUnp#d 80%,P5v],7ʻ:d=~,@뻻zC]ǧOC32 =0ZdX̿?>ٰTK ͻV!? pؠ߾c;GC\Z'ljb2{MRArʥ:4xIX-5MeFRt>q\h!V<?{.|ehv/VBLn&.c< m;=8O? h< oi.h ~zϻ 'b߱$OI !LuE٧4x O|[ TL`9cc~eGD({?˱rP|"%SMP$p[,eD'U8wشvhYV|"%SDvuW+tREbPJt"QIx iOEQ*jo\Dd;MIPbPFtbQED(Pk-ze[hLš*i?SbbUx#] C\ M#:aN*I hʲ0\\I2#]K jH/7!.' \H7t+u/Jqa;J8V|AdV"1e1zU9 єY} ۆZ\2wK-‡tXk}|,u|9*kFrΕP4x5: ނX7| )uK R"P^ n/1f1M{QDoބx0w*U 9c۫/fr3ۛoB%Rݴo<3rş^|jLu?mu!12bx .In/1Q]p:$(v#w"%R9﨔F|Ӄ\zQiѽ%˦ !߸Ȕ ܂>\ +yn'@u ="^"`KNZ PY_$n,wǙǗbٵWN `+\cGJ7%PthM2.`+D|RVHoSjP{w kr N{uޛJv&8Z+/6g|1{3I_f-C .v{17&͒|ǁq&,~ևjd<<)ɼYTJšЈPSAr}F8#FtbU0Gӫ⣯;l2%;i A&D.t?+tKe8pIr>2GztI ƠupiG03ҜoT߅7ʀ]\~B?z3>*6=jR O,ѻhl)q-{ߝu?Z|gW-FG? nIr88[u^?:x+TJp(FRsI:v~u.nP׊?- >f? I$xusJg9b`hpks`%q~ܤFr!޾=Gù?vYqr?ff9#5)VG1%fla3ly6Y<Sꘂ@k.Ťn}x:of0PF=۰Oǧy>|a~ű`$krnQrzu?7bh˫8Z%RqV[} m:FDAe;3Av `3dq%iiЋ/";*«֨3hPC/3lq5)?cĻ\Yy[| PŮs D@hW\ցD=1(ZpMQq' &YꟶƇI?ڢ&#фV]`ta|TKqW7}vcYJ={ >֋?0?CaI`Rf% I>be!F'a)]J6 (~ႰW*-E+ Wʝ!6# {]kLC(F'AJ<y&D6-;G?ڂR B2ns8hӚjz?)y&-NfL / BX,¿%]nI[etYL LQa1@ppN$sn {(H1n{ʼ?x> aM:K8)c;&Ao'b;g$/a7gZ9FX #ӹz-S<Ñ\?9Wb3NݻR z8^n뇔[,#\\%c'˅Rk1T}rhJG_TS#ŏW/WkqS%0PY'BEQ}mg+8kD"HZ0! `HQ+%v6D1 Ll&J=S{2Z-[Ej@}ՀNyU4+ ^TtP*$y eCQ;ހ~Q/^d%uͳD ǑWlW )H%륭%P59]Ջ|}=GR]|. Oa*EouIi;s% ֭RWRYoߥ v7W~2Y<3T|"w)>c%ytsG߳r4uWi94-ХO+PT߭z9oArE?k W7IeòD4䅫hN FzoT)-:&xpr'[@Z.4䅫h{b<uc꫏hr1:hbݎ Xyɨo=к֭ y*Z$C4Rb9bXF AKK/y,L2U^+ 2&"0!3Ϲh 8&Pm08@A#cddbe>|G?hM=D +H 豓=5^^L\HBOnCJ:hc5M^< *Ac3Ki" p 07Bri=Dkdp}9x;dDq%BHD> ɷJjd=?bnSP2,Rk/.PRww4~i/0=4UWZGz39ATKOkT(GmKW#Ғސn8ՍFrR ʾ/cX1{u_C_^ g]/j|uvrQ-0 ͯmzPQ@= nEfx5.||W*1'?@ϓݾ=\ٞQM4ƽ3UdqzS-a{ Iؾ_IہK' CS(lLtp 04R1Y : V_'T6 1睃ǽ Y~٘#19+8d3O5<}:ǕZ1ѳRu+@'̉̉s11@XIޅs'ɢD;ߦPBg}>*_] ʝT@FQI +N$HۈtX)b|p$qH~o#t U]k/#Hf$SGy9a֍YV%CQ;ni,3f Ʋe~arV[\ܒ")r|3.L1Z%ZFɉf&NEvHz##QܫeܽE3o a!@%<*J=/D5#+l5Ũ.(G9fKrb R p8Zx(2!.0hРk2/МN+x;ۙ!Tߚ-P8>J jBLDUV9?gOj8ü"|~ 1Kͩvә8];<?ac0+ (5K<) ^da]`U cm qYG0U/_S;-p%.+[RsQW5J`fwb|Rc pPpU'I + AiD|ﭸPVu3 B3FZӓ~a;o-ΟU\-P0' `Q8   u"hHpSp&8pAJ,8:ٌZ"UJ|i$[96J cjxO*‡c`di{9"c~OjYW=a餟&=I{ȟjЅp))FoťGnQg4nǜg)(ֵn]h W ]WHc,. \URԵ}S'|C[;]_ 2;\|k^1CNWÿ>:u԰Jq Qݭ#LK~K|7]؎ݽAO pݫ.[߽rPY+[ "o@r񑷨G-B@L󑑷P+Y{~p| )5OӜ\{v8RR8lTL;c5i!Ib5*e^ezziUK;jv@%1Ɉ.W;`uT<ˈ9bKg-K' 4Dl@!X1Q+y{~p<4C\hB:4nnD ;ڿ6Ƥ_/cG?J5Ewp,ĐSt?Bsck+u/ Pn.ȐZnXk26DH:Q=;,`KߢǕtbʉ85$\f#D.^Fɨ+LS( j%4&Bfc#fc#u.ͬZK1ce bbS:@q%+-ݻ~%{+`6w{ym1#}4W%[w铭6g2C1api |Olb1y[f *ϼ'2^HM3%#OV Ky{VL*u8.= NbSe'hv)1J\X 9 #{!z( mtǙE-:3Vt }@.vJ"`Wca[".w0|inj,rsD[V bN[ ,<rʦ$B@Hxy&qG̓*hƁZ!!kIhcF.m0nv0̒F)Ҫ%=7a:Us,ic~{5@KP{DtchmO;ᘹGCMGqk껻̊/j p@hx38Q;bۻ<%-F/C>^m_Mfn8NWvwGiiPW?x6Zބ5ra{S Il~m|V7z{K|p-EkaI@^q%k8vZSSް).ΧEc?74yS;3~v>HlKk7 .E:2>;Ѐ-z*ZI>04kIIJu W;.zBH5={[2~V{ 3Q( o[ׂI~3gdk[RVPqV"?:~D>$ȹZvNkyFSk0>UJzKRJ癧2g8 !y (5O:_N5f TwMU<5R#r ɸ 3@yΩ& Fq@ >rYuf^KYRCǘɠs0"X %3WP'9eq)j$Ll.Ƕ1(8g#"չG)* $ 7u$kM%mREmL]nRDe6 fSR/8ٮAlftinΑ1TpDe{owHt|UŅI;q,6^\V EͣV$5s9٬wνΘ\~XW6QGUEOqFxMr0rPk*(Tcx I;UyB.IVD=jJ7JA6]LEuVsy))b ^Ai3) y'?fL} toX/X/k یVL92s̩3\qՎq)Dyf$gџ%/_/o㸪- |(@7qėšӲ_hCWG׷[GbpKR"+Bimi<"p=r%f{+5s8}uK*xF͖da %:^bB5},yXԣyɛ/DXB< \ 3~ yud.j˳QMR~r?<+ǽ>~iQ_539x93#:+J7IBDdӥ긣LbshSݑʭ1r6PyaE u :F񯽱BJ7W+%4z֥`b>R=xskxU[3f!^s$Ń`[]!^zHs[CzXb}$ὡS\lEn ^Ltr=h,w7׷_n,~dz>y;hg9C3 f筞ލ76t/4t}|= X446^}@YFQm3ZoGNbq~ѤI )% K+M!ى”wywp`b[:q]gƝo]no7l}lcsS_51Z};&Gd BmpZu ̕;ĄFGnP`QXY?^KyXBvc|D}8hBgG d(I• E~S';nC4G8ؠ  bO,':D{Jq2<ͳrSFvrBqUL[K*7$w#Ghi8Gh 2JVo&MPwʫnn02e-ʃ#A3N@aq+ c>)j)PG@š$֑.bm#c3zfor5p-C㭋BrT4B) HLw<ڻJŭ05⯩/ ' Е:#<&(嵿~ }3 n9w>! )؄ԑWDo_|p)HGS <~WNutubs!h.bUaYh}|5K,-x:Q i#E& [hGy4;\݄]@57Z9I:ӗ͆6I&%;]l(%UV c}ܢshv}qts I&AJ,h @%syqTx*)ݰfFf8>n5dM'V/0f>9e+GG.yB;'e_tD;IsRn\Yg[KPYV -cP&V չ5{Jƍv@cAP`V M,p~`$A =n;P_5AyGk 'Aȼ."(ff[oŸb4oi(H;>Sut5'$6y#j|3!HL"9{N"es9rvGjXUyBUme{{,1jeu(B*\ DY?cf@*"d{ e@i]E7UGȖЊ!#OmQrs?W77j CEMmRi4/:) js)]squJJɱA 'H)c._# ԛqH!6@rr4YΩθ7!W"p[Z[BNJ;"5V~/|֟nl_PSL.! s2R")oU*ɧZ4#qRZgzUɁtVl]jf@ښQ/>PK:t x('jDAڠ-lX~u1jrpA*GiTIh 8hbUqY B1xɍX־R9UzԩSVYvnb\+ʩewHp(޲%4n4@_ve%R.)A$o5L"z,Bs]{tl)r[@!.9t9زc3bnD1̑紳d#twϏĔB6pv1"-( 07JhXe@{fz~1F ҽA"b5yG-荩BpB`f>u 1?>b &V5wǮ}9pt3_/k/SWs-sEC#ֳ>ȭseSpK V [z2^3*rTQDEmadY5&=^0Zϵ=W 1ow5M` CPx16}Xy p?KYtH~,>&(A,AWDF% bϗSѺs1䦮7@˸GẀ6'_xGٌt1tƛ,]!M{SKu(y??)quHڨe`ZV_),quyyvؘ-.T j'J%-Զ #f~!jE=?&D5W6v))I.=yQŕJ44KRV(U,"|jBSih V\JY:0 ʫrcH@Bak ɆJ{J$Xv^@`•D,(*e$zF:kc|p$Z %:jG植k[\Sz,n FBKnpL8IyX`GG IdB:MU* 2GbBŶ`΋1@}^4/1?y\=z|oϗ|͗ޒKbJ|%Ri>~AD/t'L~黋I9k?hG~ +ʺXL˩[u:[Ģ[NHJqKc0l IՁFX9)5I0u|2\Vx;wezoURy̡T77=yTjETk ֚W dIb^_.EdkGRD%R)2\6~Il Eu:ha *y5.+aE)֦ymKk?Vfl,|M|h'X )MM%&@34`fIhi>lu:NpѢ/ w1_$1gĐ%P['_`cEdMb"ؿ{0߻-bW߯a7+oj=Tmo& ꄒPf)487TcG寿hJ!ѵ~3jn`f9lC9Քsѹ"vP D7+w6T(zBWtR%EV CuL骫qGw" juXHXiV3C$J[_jA0 Jzb E3 @l&kA0[znY"N,A S2iQ xB2LN#Hm :%>05 cQ7ޙRRuH"_)F:PlD3DIe8\mV ʕB & &a3F Rm[19M }`WgX#I7 kD6؍ 0[c 5r!C_Ы )-DcF UȠ0a$a!B,0%i,@=[1)2d-h5zU4(- (2+(,Iid18Uz sZX Ƈ,8+92fp h ꩅ.)4c!8a+SlkͰ̡Y+q^4+DoK*<~g4hQ uF]P ,Pcm30sG.Ͻe[mҊ1HEL+RTӊ(78Q)J q3LI-X&m)| ͰtcO3`Yc$;`w4jloac=%Х- K%,i+8 Ċ <2řv& *Xr\8TQfgq5鴔 h@IATF:fvNm0`J {hO)QMM|;pdR *8H HJ Oia(bh8敏=k#$ܩÚ ap jnBtm=XjX@Dc!Ws1)6qT=DR8BlAj49(9Ͽz!1e3u}J%ca]GޯRTp =7z$Ib%aP7k:(,#KJ`sbn|KoỈIouշSśI\"Z"`mP |5>Z_:[y4N 3wȇH5䖒WKyfd>6A@9mMDP)ya 3v 9D\)oaRsؖObB))v$,mY M lcBrʴ]Uͬ{f2fYk-e֝hmbӉV|鐧<@G晐J(Mא0+Vw1 4Ex@'i@Pjfw^u1ش(u1ؤƇQZEg6baՀ읅tMcy,Yx,T.졃|o=z'5>-Gc-=6Gۛ[]|y֓1V/+'?_绮ɘ\,w^O^${vYBRS+Lo•PzvI3%*8CL蜳8VG/os@VxW?,W.]Qd^lkƚ|L |Q`\B; 9X2Y^ڞO藲i J\py%@} _ {ϮP{X>S~Y${IRF30+S:z0SINݛƃj4afZ Z$߁Zz2oј}rS|#ռ o)zŁ16z=^Ok#b8%9鐧j@G晐6vK$S@<rz"I$rL:$> %1=BmMQ%pV#b`@cYx,Tb@VrcHYx,dνMmm ENjm '+$]0zJ}BXf= &Aʇh;dZ~'U8"5=dzI=sT2>a7޶$$l{'8޿7pTJ3u5RQM`^g~r*{8M{-'lHqIIǠ5JPLCh 0 #&96-OeGFYɘG\5)9;/IN=,1qMCa$p3$yb=i2vF8}f䀶*>Cid1tl#&8E\>vR*VUĚR %1X1i#6 Wm[/1S2jZLevlB<n~U?IxO2PHD Nc#FkpIL,:#PHZ㬗Nso8IU!5\IXKp2&Ǎs8X `pċcFpAcIT'E%7N`}YN;ӻ"+QL>v*̸0aG{&Y#f0_,v-ٴxz&;jbzKQk8FlDaU+^Vr_Av~SY8rDUsT]N˩[>/t5RaJY{H"ZP"X1 5SvVm5)-XRSL N0TNT޳_,A4'"\Q$hݠֺ%:e`GזANۇq"To7 k囔JQ&"JDZ fH* Y\6M\ޅ{q;~v㱍lW:3ΌHupspR, U'ЉХ>junIwND嬿ƞITVwOm[Y9*xL>tNޞtK; rN'.(٢$RkCjxv,NLkbk^ǚpz!CX;#Y]"VM5<:d8cUWOf1˻h2մppZK|Ƶ$7&жU ΓxȆZ-`>3ggvc"< Y6O-bو0Ӛb̠ˈNthp_o.=.u_hbDLZʲ>8Zr7L4Tvd`TFL%&~̕kA cԐau kZ$m1CI(K5eYR7q#BT"ÑʺyͿMŕ_m~WuEBёiqK[WC[JϞ(VOf1RɺhT3KtIBdé,K̈!YƠݸ]W{O=֟ N@Ii/"`Rfq+Q^E)іuhi(7 +6HN K3ȇ*y,1G,$)tl*y.& )9d0kyм2n0f}j*pLዻiB)S*&*yn|El,(CJ9Y/@UDí^W{aWOŃԩ1RKKCXi0G6#5'A,)Ţb)Bֈt7rd /^ 7{2I6VwK3,!YOMvNte͵W'.jUYkpO_~û뫪(:.qJuu,=x6Z.zrA5UU3ȡ4R9gPXґzX0:W_ŘP_:JSYSQjPWz h(ƮpnA~\Ow.PEsvA]Z}toQW&Ig#íԒgB6cVa¥֥Rrs[--N-$v2:SWTDG Q* 7Rޠ1 .dœTr4(:sB8c2J%e?c|g]K-mTu~Y`5辝sAJ0Ld,`$aƵJ% 6SZDCkA0pt%LWɖJEKtF\Zy MvaDL/b-Aъ-?OPV/rQ &-S~|鯷;1ߟ]EldCF{.FJ.IRiɧKJi+U߾0mJtNZ J8R2q-DhHF*3%UYi?`.g>Y,.70_x+&&/ wŭ!U醈S4?]e~z嘼>}q7a?*]O(kU[/xȎ5*En{b׊_re=+SpU,䉛hMQ5b'w tbŻ }K5JB-,jwKa!ODlJѓs$3zX BL'] ZH-qwKa!ODl0|z߻qGEb41FSv5p-ahBlSi|6kX,.?-3UґҌrd2J1XK+ҩRS_)jZӐ2,{ݺJ?MAc H J$Pz-9Pi/tfelO>O!PtQ9}W,i.jG Ar!LS1|x W˩`} AuG"W!2焭mJ "Q&TWJǷ?Fx—xdXe_?8uϻ O.ec_r&1vhE%ך`Xb.]O6[--Eꖿ[MlJbWZڎN̝/P_NnzĖu}4~#`~`7se *T:䷈1Oj4@ٔc%)JTdzd͐Qܝ?y<6?|).=8 F[w^=?nn-eNՔd4tpO@.阩t9)ɔ܅3SP!vh ay8p,?5Y~ñ8-Gw#`࿇+3HA{{J_+R!hWOy W"i=O7'jV{$繥RKԇ!@?݉į<TCn%8 yF*8D *4RD&g]t V+&Jv\  ٌu#K.jKVZZ08a!R(W"&1S_ $jQWTJlPadQY]1#- 5G<7ev"6F -A%f!E2*Nm78wBƈ༐PFH)CK0ӏ -G Ҹp 24D:ΐ!]*YrF Khi-"" bk)/ʋ2 YR˅xCuhSr,)NATu tߎ@S{nf t%!po:X]Hfd6a,1)Jlwօ6Ƽ 8/gmja3 mlYro QdL`\a8A`nyh f/_ LQN3ca;œ7J!̖u qڷU^Q44;Еa6- A]Nuw-"9!x TNsΩmG&@)cdjVu!X=LHAx ĎM~>7QJ?̵-2Ql@侑PĀ3 G2PIj.OH(ZGgli"Xě rYqm|v}Gt0U.|?6ݜ<4-+Nf4X2"LTL4+BBŨ bAq̔к0}hzwm)Do$4M u->N㽞//K[qUZpVj:|V|j}?eS(Bv`Gpj0ab]2t,0rPIhN<ZX8se J  OuVzy ^?߃;bǞosͺ鉻[Q7^8Xj7_`RIj=8xͫ @}M2'DYh_^@X]rMߡSE&?!%S!]Lfv:no&%4`l=E0G f-߱@c,$n{F0 Zx=M #P1Rndxc?\_x?ʖ[`qs=/52_Da\Sc|25Ct1") ]g$w;F ҽA^kv{JnwT=ԸQ!C A,^n&'2E@dw)J%[` n mmN}ۧogw.Wjſ&hfwX˜}+,nU^ӃkYC|@%FzlBŦ|O!BJJQnI{P ."vBE( }z!Ĩ[á6f=cK֣AؐH;g=65*/)BH UF0 qrq>?^wXfATԶ'*0T^ЪWޮCkN )~wYbU4ɵ$޾ׅ[b19Axp{Ł׾[F<k]+];ed1D޵;>>Z¢]Gx D9jXe,fK^"8[5Y$@I8]X[V%=jx:Z86ͱ7בR1)=nDA gڤLPՁh(lԵ :B%*u NPiFt$qަb)4+KS8Yl$`CZ9d0.57%gX1r,Jqq%"x'lԂ+f]>ߐ7o~> 7;,x%fjr̀>nU[ݫ:FLUJ77We|-3Wى;CA`⫎izu5qCb+Um &֝氕vsm[1?t1`zgk%t [s5t$D.xO)3і|:=5>zzpk#5Fm6uwoy9lR&D䷒pӆ3iTH}TBXo{4h+8䷈WLD[)b0kY|uJDv[Uzۓ`pi\j7{R+b!U _k]]㼮2`EMJԓ9W/<٫ɿ˯'[ ́[z1~োZ~ZV5\b߱wo͋hhMoiw>Y9j{}c0;.`"lUk_xzV7?˚*WY7m!$7w/hb11wtn#1N y*-U2J NC&lV^ cͻŤ7NMe`l4-~4 ]KnW"84xO7x\lDs5seoO)V#Q v .MH($&e!6L;[i %(P=[ eHkƿə̊{iYWL"nmdOxa{({H+pbT:eh0nhMqwK/f!$n C= r֑,K/\w,;)|ogıILpH RkU Lk*3 S0"{[bnc*kUû>/;j˜`0d4(DcMP%2Jl(BWĎ jWd X^~ Vu, DאɲaSA)jtc*˄ƼVS`7Yn䔻nTr7jSr_S6n,84$! uJIppFD>) 0yPH"\"e!QX*Rç%Fa-C~e63pUh?ww*o!?md۷N`>Fd-ҪJ̕jB}UĖX4(*z\S>5 kRd↑ c>p #LgMC+|kb4H"K?Łxb0D7d4mm8dqҐM!Qc;F,`Jh8=PmMW>$eX4ksbR?ڧFN"⦠%Sp{ Iv.].FʼnEH3ّ=L).v>Q -yvCuj/x\JC_4Q-(У姡Ƶd|qLNC?U^;Ϟwfcc5}<ٜp25Mk;_PGK/.-/.ݽٛoVC΀yE@gamb;`n$Ӭ@j@~/NId28g<ƺ{q2hYut]'\nAOc_sa27#~fdZNP[Pgb(|k&`ybn9yqb2>"+.هfW^+vV-QHuysxٖgv*)^֯4hd4\6%)ZLu?8CʩkVs :2Fo4 ;/C/ƙSY7$%'u_]}^UЄ47χXگڮH+hkz PyсXfU]L ),6016n1鍧v0{tK%뛻~4 ]&_y5׽]Pq~   ^y^$ľI)T9ڒs:~`,/oFHG~ wmNXofA<l$WLf`7w3]~W kjxp$XB!ZP:ctLqT$'Ù CɎ]ffyzU_t[x3vqb ;LWqoc br{ݛo[jr(|GZK;?ymc)R!W `rixE XY)S"2Q|J*|[&|>f|Uuo>v|V},$q"%1aPJ(,0JPj9Nt(P̸fk%CmK@ӐP F!$rD`E$JH\ >0e=ycQkJY!Q& E, 1Bhǚ}B") 1ֵ%E8 J4Khe Jf$L4H<8F ]>Fb`-Y~?OE~/; w}{%?C޼^<0>4~y ""@$p?@O`Mg%o!a})ze347`;%X?,F#x #ے A5WtlO{&±C\#`-jtܣ= ^{u'_Gs`+1VD2Z"IKkâUmz#n[[\ٍ .ՑooP82KfK:Y&'ˁa9 s; NP|>T,-S6acq`7 v`$g/Y_k2::4Dӹ{fO9$ZT0,l\a~G|.;R ǡros? CdVEXInLmX;ef}AҧU_>yI+I 2`)10ã8i-v|>0߸ylS& 'ަyCyQ,kdjvƺD-/fQEֱneK S04x 5W={PK9Jŗ+p}~IѝQОBg%R,$ }t+CLbE]F4:1QB1fJIeL(e'q V<;u An,/RrJv]"LjJ$M6 %BCxDZG xJMD++*%:Y߸fƖut~$<0t1zւ@vK xS(MTqۜe$g7I&gg4M Q5$TK\*Ejbg4,jtVH.Sr Tmí ڶQ2$퇔LT,B,e$4Vj̱e[.m^-4pHhz1XC -f3 @Պ16#f3GБDeeټx\a;zMcEG==CThzl^q&tz4,S-5yqume]$/&%ogRTumdG]mnho١{; iZtv}oh2nFWr$zqK~7BX ʭGv>j*h*/Gjq ns\p0F9s) ɳ=coϢ]Az恑{ w \^_gwrk&؇O-v eй$-Hƌ:Mc&pPm02-yyػ/Ht1%tҽ>]*8vȦ6luIXjA_1VB։ݾ|+貫j"j6fZ+>'&.쇯מ:g d#Ϭay!w9n;~i3v47jbbΞY; 9 3>^U+yv:5С{ gCG-.T޿gRCs><EvCxg]᭣AΈE!," ix@{M@fך& Mm{`#?S+MkSF*m%*?.Asňꍏ͝8{ яIrܙR/3\ 9@tLgmメ8Z*^c8A>8qU }F;G[4܀wZބT3"{m ӄZaxS4 (#&t*Uŵ0{'1^.0o}?o_OOjş~9O;CKj̀NNay  { CE!h43R: HC xg$lFga`nRo,yP Jeİya)7KF9vߪǟo_IrOdDJWxu}>EQ՟d]Z,ERZ,EΊi|v $e`9N\3cegQ'HӒ'䯓ɛ>]T艿SIۉN&ыIʹ0Фb9d>&cx?֓d%WIx0z>y62BpV7kB3nI e^[XapPa50MKr߾>UV1} }E s{gpF rj'P2bnQ $~- 񔴁J]IQ[[ׂl3ozVLZ6F[ $^ȉ`sCD-qDQ\AKLj\ UzxXN:r'Q+'rA Z-urͅʙĀ5*NtPc ɰu"Xs+6t تc:n`e@ IKP$!P;;.JiõQ$.8ˣN*5uXXjc2XF2isPмQ+ġ9T^ø*Aѹ:\eP <G#Q]hPClВ8[F7R|#.5C$e߯2}znVK<~!Nْ,E@y}_g'&3ƕ[A篧?ί| :4[7!d_;zmY;WPxsy'Kh[P"' UmǼt̴4d5\ Z ŨИdepw޳=K޳=+{eps1芜4\<<\du=8QGt!% U\)hO W +RǕHk%'r"~7ȄYpE;}sbqRIK+iawbV `ȑu״3axh9ɋҧW€DR!I=ЌA'd"A 2{~LȈcFB F ). jq,W.BF}*Y&<|I#$F&(4e|,6W)n/8v޴n@/SV>(m)ؚ\b:Y}:EIiR; :OAM( {ZwЫK"sG gW( %W5 8X SZŶe~\=;:O sʫ,:7rLV.X{+j=Ť^YRD@oH?k/>-V9 9PJwj!Q3JLhe0_ ZRT[y\}JcYhWz' ZQL0Wԅi!h篯, 8D FWs!cXvl-p&o//]w+xd5·='[q ;0Jn=<180] +9W\,')ޣ1G蛣; |pMV/-\RQ*Gy`מM( a4LI|}N٠ F3< @< UYpF%Za]fƐBuUp`8NxKe 5hӕ ބ<Wc@ѴOT ʸ[.ٽ,-3+8ifM~I^ LJIS|z:-៏i>ˮN5֫wa쎼@(*Oh aBՑhN ȝ:Q1_PV3JPG"]4iW- BF<ڐ\REڑt2KB0 l^@12(J4Z0kSƚ#```QA)?*4(vWW`hTTZj"I:P^.'D<`:@>:##20D *6K~T(*ܨbז~)]hԂlڥ,{°vo٠BnZcY"ߍ8DޤA%?K|ӻIA~q\3JP܌*PiA$J%C34;xE7),n"nhO`}`,.-)$SAfD'8T{Mx;!t;S-ӵw%Y-ƨJc((kGʪtCv1hM rv q&ځB C$cMAqyZhFnޤ#UU]]N.²rjLz?06m F׷^q"0HmT 0A {BȾhY*5qi*4 D (I=5i%qgV W9M4r%*HIߡ[:R\-MHFU"yr KhxZԉg\,|Xwi-mm SߛK(h ß  xeg7әn]ze+jI*BoF^XxnqRGleS$~w ouufbW^-7+Njzݮ4٠ {B^0s{ |?]JTh/&㉔IsKXYJd޴b}#R#WEnp%x`Gh.Tru\Sw|q]8FlZg>@ќYt^;5K] %b"Ԩ\i/#MҀ}76R)=S (tJ\숚dWn8 R͘,XwK͋b_jdm$ꞲeUxwGM^ϻeUAi#M+CsYGELZ8ĺIZhbPGtĺ ӷ<-\ ۷uku[ELI=Id&k֭-uDMۀ<}Sn_uBB^)E8vxXti֭-uDMۀ<}SVoEZ.$䅋2cö&(]kIWYjkr֞pw A2Cb7$ * VzSٍCzZp!/j@n^X ^ãE\ǘyR௘cpj_Ңx%ד`juѢ5?{Fn0/INۼ_ aɞ]%fnx'?QRnHaX&U,Vw!q)Ӷr 3>oM][1WXŇeFtw){!MQN)wvm@mGi7O=o*Q=/ s>b;پw/՜~W&^)z7VAx8DDVpd|V4G9YǁlilAlV#C|N NիW|zU 6'N)O?Ǜ6ED}Zdb\b8U'_ 6a@ZBBLp&! IQ8:Kd`!w I)R?gM("O/sAW[&8y6KF/V- EQE*M+5D{aÔDaTTZkPR(O^sZQ3) G񸕷;6FB@:>צ{NH6:"`kƝ /bs6>['LVm$14U4H$:*@ۻ50_IVVgw?/~ {o/~Qoʷ ȫU!ȃ;} `$l g -(#SF;K{=14U4H896#7fC]& }g@yk^@ EɺFbyx,6{|unu1Rtgp"4ҭ\y&bGP(XLDj$024GAHR4/3>BIGygk:r<{FPvy}*w)H:*Y`|V/7)c2ۭ<:E_$hXāYvlOK6|ֲ\;RM%)SB?f߽؝f90U=V?cw`@rHS˲BTL4  &uc+MTg`Ҹ,+ tU1XF!G`Cak3U&x>JOCi wn 1.7|-.$/]~;hqOAKr'2? 6%Mc q!e1yVTҬ*50`um6DT h RFO$$@1l`=  pi^0[6,.nLf-u{pA^b@> F*+pAKTb &u+ 9U# Jc *00:GBj\i*64ƺ’sͨQBHBYq@Cv"l$q-ƧKfv[v`XYs.Z?6quYl~Z Ch>_ O?<=ZtK/D?<~<0_/W7S75/^aǧkh0bg;jj% ??ޚmf'#(!?@cE2T/8#aZ|"0FFh3]f &BTFc64ƫR֘QiaC%X,'dHUmR``8$3 B^ɦ}uYk_Eϭe CZ0kP7<:#!NcuC=[IѪ_ wG9CNi2ȩ!ӔE 7䶏0 rc BtؕP6ucz%QhL?GVi=bnD0yٞcZSLtl9-۽cH7ӥS5@b4zTG#?׽׷{r:o!3GIJ},e,ǟ>i3Ϸ7fc<@Vb.ע `He3b[D8/#"&[Tܸ L39NB$:0aplAxo-b(9,)Wyűq&[;e,Hr<5IԘx(#uHrPP0q0`~I `^,rJ,'>Ӑkj, N )Q@OGp+6)&pBLxtb‹yMx:"LH~gE N .AsնVXsӓk|_O g쮴źoۭ wsgn0_vps^o]`ĴOzKLWd~8kU/+݃}?߻mW^&]]=_㇥iiM%qm`wA_1 F釋`}NB4ɬdRcsxA{r,Uɞ>岜E$ 9q )*O]ݸLb:mY&L4/n2=Wr*SN98}>LKp׭k&ZCnbB[Wy\Ē j,xyAJ$1TE\C[lw:9>-,SL&cʤS-$%1^׻D .W5z ={ Ԉ<="הPS$">Ne]J>~!'!:E:EF\wW7Կpu GuB賺MR)h[O- mc?IL`z|g/ ; =!L8zEoNGpH%Zld#V݁'}3Pv6$@ KtJrHa lj˟?~_GlG?:7*Z$lZzOGٸݞ;=)LoH0RONR2c0q<>bgmRR0-jEA0D:k+=ӽNN0|>hFdD z N#?}G aQ/ya ,F@DsRqX>tK8 ءzZ8E2AQ`9(nd^O0&'ϔ@B(8vs譀lhhyn.*˂dٗF9B)eCDrfHe%An}ץSk3q?j>_v8Ǎ2?d.G6%ȷ, JU?xaF֐3 rz<s c^񔻵X${]d_1VT)־5gX1bRNзlzy3Q1ʵh RIj^Ĥ\FR[$vƆFoDGAWR 喐 +hLsxyWt%_E79  iB؈0^U W&#!2 /[i`SIu~M{ -B9Ro9ΉJp6v9hzU/̶xpWŃn}G>pZZiq"'{4\bz?Pf}%eQp#$&pfBj HuJ։ dicrj4<䴙)FVOTΧC3QWi2ƒXsuax:fj G dX}q;2q"ehB8|̕vNPFg":N C^!Wz aD`8wvnl`N 瑯P& #ɢp68)}[ }2ZiF1l9V=&#聝9f#ͯBrkEp˅,w! 4>`irDZ4{4~4aˀKW>rRI|o&gR2).,BQפ-ޖL u#K3Qmgz8_!r6PKMl'89'Fl^}ДV E:r[G3_UWWTY7:|PtӃ&EL'3oF&6\°D GQJ{ Aߙ<`&(^\\fFCw5Ufr ހ-L0'F>D?B^D|;0xSL޵yNp/ KyŪ;Şty~}UPY,*+ݦ6 !D邵:+H"ӊ*J/qopސ& g BKdX xyfg-n^4zv#9Tn?eyŘ,(-h Rd$崣Vȉ@=GԢ6 B3ؽ-* QkO8V M|Kp `|Z49\KRUiVG'Bmh%e;@cC?@8"BK %ƒBY4Ge T kGK imKFW Tv%Ԯ&j݆\FԬڈhۛX)knTY;FW+ۀ&NK[iea MZ sXE%#v#kФ4і3 -k.}S;.Ck+ޘAfwO]LcB/c:vc:vPs~UWo\-Ukp!(' 65 k4p  O^$=K*,9-D CjO:ȁCJHhHJ8eLyBFr3Yս8uxnjT rFV.Ę^b r"D>)µ#T eNsߺ-K¸?B+MR>0^$x&cOnaYMPbUV(.#fZRkY`W3L'r%) DQS,:6&V uZo° ͈iEI!'j,8oq 1l.QL0Pb4v$A 0w~I(c g$j (p~-m7eyUD/>'EE+ EWV,lMFYQFyp% ߋTY ][oqv3=S~!1UE1Us :sO($LRk<$51Nt:q[pY`?ݍ3[~Uol[V…BC=N7i2uI?zWю {Qo{W'Ou >{db=ߥ 4Ylv bx nӼ\,(dTLl4$J}kPhx::)b8satF8jC&Iy8V`H'SӦ\ʆ#G.jZ\+;6t"N[M$KN@j ДI\/Bp:b<^T%1)bO~{Ԝc‾}F-{YHgkD =; Uf;,b"tciy:B46cP`]ʼM#kVJcv2Gւ@J{魤Õ|YjfL~n֍TTbf|J& S`!o>Da\ IMECP8&$߫g@=Hz< t&`.Qn2Ppg=Ɔd{]lτg"=*rI 7Gtg.XOZV|B> SW;8ꊻVp&K*'?#R~&ht^VG.i5:J^-h^6`1Dۛפ˼{7\~Gݳh@ aU9S ֳ#"_֒Yv^MRhղ*:. qzU?bB8:6}ʼn_52K"~#* "v+:[&v3S_m7P9i %0C>T⣊юR3è%P\dBېUFhԪ>@C˵DTǿ~M!'WiM;6^v>i0}!bdq=2-!ܱ F=P z+A75eGd+0~sWLMXO> ~sF g"˱Qؤ:iK%%xƹIqq.]]'.7߅[͟*rUa,uK! T։thTcKOBh }MϷlچyk2kTta: ߁6#k}Ju5܆1_:ϭX%%a(?w%TGLK Ev=mE,"^LA*|>* 3NRx#iUQ `sJ5cщ K8PnpJϥ}$WMՃR{Yc l`ꌐi8'k̽'k16 o')J+~n>쨏 \n,0D!3+Z>3A`=Ĥ|dV TZq0g9iy04Etkm0+ GXEs˥́qT M.s6g6Vl@?A؁lc|c=6fyn UEA,.?,s#Y6sJ9EY~1h NGR'(qtwOLQN%0s"3i@kepA@'3SA9~#&Ἣ$5,,˄M8x%ؙ@Ճksd'X#J{dGXLWKkA5y>P5`#x[Zr " l>ok,56yoՙBKչtAuRgqtwSdπ"g+w㖛\oə$Wf_г˿6˱pi@\ 4]S HQ3H}-GΝۈ m+]UҜQ[:ʟ$tzaĿkt8g.z%Qe +_ۻ6|Q)lZWBpws<^cǏ-\|ݛa= 4j  !. @p@nGֿbBC \mX c2r o3JL`rHɔ!!%Bbyr.{Jx֎ ^ϙNcvgZd L'VN$Ar`uhwhS^$Tb֨oUge쥗8R3θ]8͖Kla_X3C՘ Ɵ?p\>%T{t|ѵrn݆Ԛ/mlt;7MCzSoa˛WC,7~V#{7~w/ĉ(w"e\Kc ԺyAg$7wqRt\?T:lr2E64̵1mo<YOz*%UZnS jp j⌇Tsj|C>s6f"ؿ\rAL) 4FE);36Kw5'\v[Nc^/Z \%qp 4(H m [Re,7J"kQGiҁD]W6,8.ˤ!`)~ZY|zo'Lǟ;&˭xej۸_a]v_\wjkoϩ8 $JGRvRW#V\-n<]Dok( {sǷ[_\!^ >[T_F<9㙑/x)9߽x-Egq.oAҥ& Br+Mf OٷA>IU+cmqT<1RB0;1Lc6YαD HBv?ƉK-;a>G`+D琧$ cM$J-& ]Dj\S>U4hrW䮊&W!eL&0T ov"cMf<`"V1ɭNmP#jۦoهL`9!T!%wRXw ev8~Rn.Xʼn2"Q%avY*)=U[X<u4|{YA 4d+c>!Q $Y@>`\rvOA@׬",XRLv r'qcX蕄#s~),$!i2BT 4@*D& i);Cz8]i?{X|YƅZpg{ -j>$s&!cJh6/%!CNqwbE oU}qp5uwOnϫIﻹn{=N_E[j20m=ϊ{~wm>҅աOm˧{bmU( @Ii>]eO BnRAC h`v_k;"?I0uo-I9zY P K)9Duun^3NQiv jny+ο *T+6,Si-skK!mSϟv KxOO+k)(=de םnUMBttR<=Mu*UNCDZ0r.:g : # Bi1Q~uŴb0*GXWa[ͯ%/]bT0)ai5>uw5MQF[fM& 䅝p. i/ Ӯ3~AU.c}"0CHuid arM<\pN!D&;) cI46GJuHro u{}+9U Z9}5q9b5d@jFN('FtYH`Ufi o7x.LXl( &_K} +S,`c1u9 {%1Kϋ^bgNsKM wvP jOx`.jK,-VmfA C-Tc8Uc_ u?BbgnXyOKQ ڂ2jՐ)~qn|h LC ]iUd8c'Fc%`xHc'hf涍aS :ΰۗ%qut%d5ҋ`K=KT!U:LL mqgY{]e[ؙrZ*ს /JGMmMKVY_iL:@j *-CTn\"j|ZiT7]:]E9urLMaFj()לL-Lqe&c>#ɰ0n xn=W_jEu!$C>ڹqrL2|3'Ǖ C) (#TNË#x<vf-qGbmdHq,|y.glI8?z;şW9?AU^1E_/']hRء7Rf[7W%r/kӽ\^yf~{ڨo:r>z3gkUΗ׎rmdSӊCɻMQ2iFvyV[㎦z6,䕛h}bHc_優\fR%zcs2G㲷oTɗ wAS޺L0ƺ?O|0!bߓ`_jph+rGisj̼ce,ƙ{,^q~{KogEzuيaw3zDX4݃EKҫNBwڿ N'Nʞ{Z2Ŕ}3۷~ IicXOPyƠݏ\ЖXAv/:QT=!0A{yAl Fh c)}WȮWL}?7hhu2|?sl|a1\E'sUt2WE'z:,, b4F2b$^c< eU0f0gYa#S7k~urԴ7q hK_o\Ŵprc*:ƫc,S5%c-4\ Mg&@DJڒdhP#hx}F#QOtPKY8ǧ2>GS稂PB&)d VMad,Z1ٻ[o~|d([H8`o/.퍽\5}PR1UC0mw>Χ B)M(`OqZEdaX\Iyy 9)3`,1*V͟h1 #!dlH&|Jsn, CK6c!fF*3*hR3NR%f,a썡 c Rq+ CY.&4c)7{*:NoW RQdKh#‹Gy?ǧߒ\ϟ ?}~Qx`xhaOƈ9hӏ. ԛ+ؿPE|83BM[3_yU3/;+d~xN(lKRX{t3 )L X^,/}Ʊ1a[I,l +H^b|{KA}Kff[}ْ4mc:#0D(50,Æ0$Rb'-כYf8q:Ɛx)9-(˜RyO1nZ 3cbk= {VIH"Zd淖B^r@7ߙXg9>^p񫽨-7omNeK˂b_ yu3ݬ*&Qp*Sy;70q4sSb/LGG 'wcok<^EЗؽFSh8]frq`VC-ql>(rybTiHlPT]r_iCy9hܞ{ٰ6p|XMɺ/ِa{:-?a$||,}-UT>Ox=dt[}bLF>~)AO>“ 4*A[k6w{Yl,%]>Ѯ.*^.d/>?Πcicjzڜk޵~>ۭR(~@UUr@kXs:&R]~8tiօ:`-)Jc']T&@'%yƒTTCNiԴd3X(SS03A:KLFxY$u(H3N D(Ku5 C٣%P"ԗZNO`aq5" $%xVicDKDp8T+pm壕{[VQIkҖɼ:ŊeHHД`$3FٜJ|0' X/uP7-3d< U~™b"&X'woQ.hLZ A q%sywL$&ivs<J2'̴& KV̨Rrh쥚kMLSAt a E1?[$;tI2%1yqXqڛW*H74z֕Y$,&>io]f"=ȆC<2LXqg"q#, zr  _mQljoPz׆2vϵ?},7wEU+KV-連蠱/ٷ"=(^|3q;ye.Rs1&Hˠ_S<6 ;x-EAR Gmb TIƜ%Dw,}3 m-8 4p+hicm?`AUӡN)B!WM~  '' jHE:]dy} jbG$j9M RrǠeEB3xHСȿމ!9yHևHI:U(1m/]N-"sV +d? Af3#y$O\ ,8Z*>5a%~9 #gLZf*Dnd& etA |T–IW4-gfRe~o̓QLLQ#4c\|*XRxwm͍JrNd/Lԩ$I6/IHkc;I6 J(@(۩"h|h4JK/Z[J!)edjOA mUsR1fw?NEt㳘/ Z/vEѵ/N qxZ/ f\%Fȕ]r"/FM2Hvd_op?';qV o~L$)!W`C) `CX{pg& /\8my C( &ݓsl-a3U{_$v=[b:K]i"i-7s֯͗s$D|{l [wq;lO{6H5 >v䇇#(W1wLpחpB%\}VbDbkQ)ta-FA?]e UG=4<5&vZw*.};gL*7޻8xXڟ_P'݁ }PJNyt^W5GH,H]:EA@hSL Ƥ۸3?MȆؕ.ktfL|5bG5>?|y_&w!%“@- sFoS$ݖV:wH Q;;fcs.mBdubw5dYJ XER$`"3R3k tAh2D(jǶMR{UTu%ҙ[XZ%[c+ݼ#I7h,I,v[smӱ*39,i[#@1%2y[p2ʗֲTj!Fin3jl|+UH31Ę48bBM cia5P,•6A`RpjJ +`0LBs+81Ξ%C̥Q RAS7Ish֍ޫhtAJXAlXrfdk)]#o2G --ˏB ^gb4i&H+ᙕ曧W"X=<{]p!h_ֿ=Y>:o%Fr)K4)9RB[r}ꮛ )A% b)FOJn u|Ҹ3p&D(Kɚ=8R{3[j:I콧vnm'k(gJ,\tj&]$Y#tKJ2/(K9kSOd fR YuJQ\H萓+ N TV:MU`22iX VNJ'z/ s%~߆1fa)q9kiӱM^,yrxgm^ RLyQ,)y.*IY!{*OuK"Q10KDz~wS24Wqɉ_Gx) ke%Z=ԕBB/=YՕdh?"Y3X_Gl2`_PV7X\QaEb'd~?5v8A#Ou:uul]7q, n1Rf2!IaA$)KdH&ܰ$1~vX}ݸEؽhcm Y4 nr$1ý}|v(^Y$ȕ6Z_ :$&5'a[$[b //Xx<$W#~DL./H7ßs=REޓZRΒu~yL̂Rt_g T/!}s=~xy ֝ϭ׏>/(#xVͰwM\!9?rĖ!F8Jf??-CW'fTwgӇ@܏3o;/zM149/xaM3mez?5XệlϜ"'ixǪkasN27DEǻkDŽkK08Û8a9þ +1b'4(gk< ,>]_E/g!,Sz|z4VD[GFpt ȉ/jMdS]t"ڢS1Ӄ)4t^'gBЪ-a?P$qe?-%s|eXiF5uQgSFaHD)C KLf({JJNiL8& "wMQPIi|( `g¥gJpŹ"Cd+ P&ϱ$y=H&D"xΨX+KsI6giuf9@diPOn)rq|A v5lBF;(,n=b|8͹Rx>~z-W+wSd翺&$BIA6ߐ?(=:Fg -}~gcD@?\3/sˢ|a7BWdf=aJ6>yI`,$,A_cK)ErE F)5z1rc^r޹CN5$l@Y3{=)82i':'+OtFJ:Jv/9 ZpF(jqt 4m̬UZţBU9 O"*w A*E9Vú{_X һFV-@K8.݇jf_jxcu(>tm4a>]*tK=fO{2Vǰ7n!6%MB116xSγuKzzM4Ħ[R m'upq'kq=")A; N1Q}CS<"T#<=Q=&ڣϾ9ZI14>& au)_uD~ 3"a!oDlJ $x7Q{B鄶Ļ0^W!ܛw wa!oD[۔RԱ$.iS Oq(q'0FH^Zs𸑌m48;NI> 33 !Q)s 3/%o*њ\S~坽?Z>qpثov ]AB,X[8B!7Zp*Xk S+Rynr0fIƫ{pyY=QVcS^>$U〴n_r?5m 17^hp3ܡS M>{_W}+>d=c:܇kzw|"w^&__{COZBTuTeL6,w iG@mFxO׿ h!ago`m}Foh) fno9%3_"ȂțAEON}=gGiDqΪ]TM9F* iuR_[ܬ"2Bp}8CVAb'W*1=bE=ef}3V61 In5V"" ϻJ,UFL"B:f~lfUn]ű\QaQwa28N|as1lin:K-[ f6}KalL *Y0QV3QEB T$!/%hM98KtOS]s%{2J $ͥR>^ +A~b远Y%[rqX<ͯ./1t +%%tvqp-!s{ƺ=ݿªl"x~FɗdZZ2;>)4Qܕ625FbE׶hK6%ֽbM<&ř:kqƤ`2v`65üZ#\~cYcZ,|?,1vL+6*Eޘ 4%-cl̔Kj2 ?rlcr`*ܫSԪ .Q92\`&45R+`g%Vp)M.J@΢k=S 4PT+S#+EJ%ޭgIvO^_s.z5mChRJGq >+xa!oD76'4"(^),x*^= :zybneHv|+n(*ǜvQ]Blq>D`%v-T!aM&Vs~G:$Wg')+c9cM E鸦*Kۡ>MT];.^b63ﯜvNG3ख़H9Eg8٘)f>ͺҒHj2472D|Uԍ㡹JB4Z-gz_i^pxk`9b7A%@j}K{2`ۦ-٦DRi,*b}\SpkCuW$t U\1E0]RAze92 sqΑR3/7:D-<פnOE((P%)8ح)+nD+x"]Po{p0!bq&;l5HLLfL2 Ke+~-ƃ}}f\мdCRΌqPXr(F*erdN8wB z'3j.;JqY i*.E̓ Ce`T@= D 2έ YDmD4D qIcRBB3 kdY @LQr"¬rQ-::5PM2ILt&5X"43APt:7EP\Js)X3O\P.:T5N0)+cL I ATJa ^,ic>XE)hr+̦HA F |J8X' ǂ ]ǰM89*aÁޜ^e,:Q3 S>kTN`_d| Py6|.,_ rH.S=En!s$N(ρuۑQzY;0-"K9l1%?Ъy?6G9Q7s9&@%mc}Zg~B(SQ6njr`sF,,NJe#nʿNѥ5*JA%ͺ)EI*6+J (0^¢3kU3@2HeZPB4jNcRK3H?øRK~RR %k{cUXwyF'%Rɐ~|[mzaw*P +ˤLN[@&vgU1=pͧ3F"ݗ[NSto<x|sûQdFnN3xjNʩ[xwa!DlCލILC11gtn希N 2 /n1,䕛h"=wSpd;"%C|A Q AJFҁ@9v,d;-N=wW܇c#Jk\:2DQ'c=sD2 S3tB0QB_,8"+L,Ec+ӇLGe /w&PUp>*^ZfA1]H]?)i1?bbPS]vۡhJh*޻(p0Fr6y[)< 2<&#V4J.rU22GŝΘI Yy4[v/J"/EE9ҋP,NXD@aebL g,Rh0Cy.B;۵6* HP-g Yy~trHfS=o0YթQJݕO_'~E.upSW  p[xiHO|vaߺdHSJ9^5/,2_t%aSc.s,'i,X/ 6! _V+J.Kb"2O%̭0%N 9G|/^hnY/~wbdlRWA uqRRWML~Y[`n # 8$~ejH}p%RR+'D\c~0LvT @|ٯU=khE7?rԕ1/NphGw8`YE4@| A+ dTS q 粨C)%+=T'YjY 'ZVSp)Xϙ%ElԜJ g"+ b֔f e5lX$&NCQ$T>䁤JfP*96;+3VMɴm f8r̓2JsRJeSdDG.1D䌞K)*"d)רd5D 'yI8HsA*% E(hʪffE&,S;*TbD %YEP%pa$%&Ts|ҴШYQ(R1_Z!hq!#z 8wksEH;I6br Dv =H2]`Ba.`ol0T0L%ٿBs[_&0'$Va@ P'eM˾"6F@XH SFc 1:ǻḱc(ט \V.Y/jn3=Abn<-)d+ >(!A~47w$\74v\XԞh?]}2>\_,Zf&<)$tU&Zc!G_^nymEtňlK͍?T*T9d#yD.QF2f֮87ʱm_G {P.@Ye_j-fD1SfYa7S%raRdf=BhKTjs1dZiZ$S$FxkZP{('ƻb(:mωϛƼom[W5֏A?=,W^r\-+x4Wp͸u;],70=L7%1KdcD̉TƦզ]=Eںkdo)(|g|O=vgXͶw~wkWl IHR"86{vΓҙf-oݽX;Ttš)KNo&u,y!DN?>-l6$7Ev3[}sÓyJIc Ũ{]*%BR)'`Hf$ *P,ӚYR*d}9H-0Q{Q~0I]enonUU{]QfYbplMua_jCTQIdU]YZz9d߽y|(V lI\ sq|SͦYƟ1 g$/tmo^#eLJ9l`Oo$X$jIF{v鏗[}_VyzMOzru7r )N&04WnN3x@SX[zMMB0ʃtV6}QMg[ CҀÜHL٤bW>|3l:Oo'þ`·úxx6 eo_Kc,z ̄ĴnVnwµŧ[}^$G+įBt fZ ZUJJLFb̨Vݘp!S.NOFg,q5g<\s;}ζ"y{,NB I-vN6!Hى}¹ V0*TK53Kb\"4joBg[ŜĸĀ~_kL/-x9TZ-祓S=@) 8NjԼq9h+%ꥇ֮}Fn~i(1й2Ab@E$ڒ8a7B^ksPס/ 6=OVUXn7u(:eX*S}^{7v&82/DH%4O%Y셿d'Kf4pINv(dV8P`?v "r%~]}|{{|E3?I3t^d{Y*rzk9lzؠ}R0ڗ-}g_K%C0Oŗ4fΉ8h_Y )Fɼ%!"͔S<ɞ/-PR%~A1&#SQr9G`Ȑw) a2KRuhMgLSݭz'މ0>;c'6 w2} Lp:~|nN$4KvG]LSۂ! M)bKOuFdotr& :@Gte0$ x[w"wlV\<8{'JN$0ؽ7$)f}ѠfIhлMqvtIv=)&+r$J7>$oJ`:LLJEJ~j܉X{0 XL.&$[s~*mc 0J>ϳ/ }<70 TP>5C_2t Wjj:K7RPuЙdEg$KSX(D't?)8Eqg'>)2|c'dT==6^0/[[;DMn|}Ȥ` `;ɀ\xXN5#ۋnl\ʳ< &U׋ {L^0)E]MO-#Ϸn˸j2f<äXyJMyNδҴIH A%0I A3-j]Y,10C!(~ե&&ݰ?Kw kXqz +7m&:qt?3 VvRTe 1TQUUBU0 `^"S4sc.T"?\v:dL{2U N'0+*gգ!gt*@ &uAթ2xKTpԚu<Т֭ 9s=\T(񆿈ʘ#2ҊHr}B SۘK.rݕj_(Ely)B\)|"d(z RiN~ُ/ߥbf3NN:){#Gk]~_]..jl6p| BJ唔cs!#qqrr"ރ*mAp8\F*b``FjBp/j8;(4Lvg]j}@g uiYIAtyĐٜKx']zF@c BЦ=_;qolݽyW(K"vT/3 }q^Bk,X$DZ7ɲ-GGD> \ߟzeoZ>M+fcK9Ņ߅-.9)/6;=,R/b۩m`K&PSe݀8UjF .U%@Wo|b !8Q2ݛ#3:Ng_)#cx0;e?\t}-s9r>L}I5w!] QbIhoc'9TubQl`VNƱw}2tY"7hn #q2Ə$^ؼY\3'7?}jo;fJnzTpѶ2'j0l֟p$6SfԤ)/2UԼtgNG1L Tn!<㴉 ԉ9:/y'%Xٛu(+]n nj>\xѳ×f)MZ<# t \z^`]/Rub ZLx;~sF/5i ݈@[;5X8&l?.!w]"W`Qz"6mV-*9' c΀"T1GgmAtbc 3…S-*' 2 FFL$:2 ' QD *Zqb Xr7X)h Fa IK`SKJ@9*Nh6?:?Lz.9eqa/{X皁) L rOa)iMLzf^U |.i;_JvԂa,oa WۻQm#؎W.)Ap`Qfsaη?sa mXnv:G{+b%X(!,öymzX3/8&ָ0)ПzoCcmVz[`5h&/V;^`T~sStń@rڊHXڹeԚ$BYF8IcmVê NIZ_@Yd;gs8EЫn/W\$u6]ih%vg0sik"h<"T |P+06ZX: ĐV(mt(ap;u0=N  :%z2UkOu[Mh%1fpSsc3M8uXǚ1'W8R8rőPт=fK屋5^ ˁV/ElO`Xq znIp7>DHw/?21a8/(Ы lby);V?=Fp}pl*{AVݻ;& _qgH? 7%=C rP&ATSm hI𢁨{c@W5xݱ/_j |Uo[Āϟj]l]CK`]N%iZdS?'#潏|y_JIBd`_#g@C,̈>ɷ X-f+x=K 6@~Bl5b^De"Xal{=;/e0o5. niA Eko|qeaOcK߼m868l \|EU B!tM,XqI%_jŊ#x1gDd8CsA䅖H{,s sD*l  POHbF‰ oY/'ZY̙G2HgC{V3D*.瀛e |_>\A`eT.z8PԯN)暟@"^+"2(F(e هu/ e{i|h#-dә[H8g#T[!z>..qwRȝ Sٵs]N+ 2XRWezї5`;$JB9].XHLIges$\(#:4wIt2ӳ] `SFSOsq辩uW[7E~97] @.cBIXys[֑,#EP%~%1!%Zx3: 9խY6xmL9C#.Kh$og^h݀aT=4{]M'3]׫ѺĸOmـJn$\y8 qYBA#ڨb/RiK\B#EU1_^P9p@hL``ө,U8XnFsKhvhiCKM7 G+'p`LQ8~˳A9Ɗt0HWQ9ѹIM90$E D9y)=~oꃳS5i8t(dedFAh3"`0H.Sb.ݟn/\^bnk~v2ȪŜ=(6jdU:K h)GᣍȮ2 F/nԚm=/6IVmj\/mzq;]F~d[M ׇv%9[݊=s^݅1s=["HW$u뵪)D̆PZ)pbe" :jSv#O'@돤nefnlPw ,˭zX/5"OCU  d)-ZyuQ=gj 3:/o11q62jZ;nNtDR[!Z*`}I߽ݚL/y5il\&J YeJƊG28M P!S)(L_Dd=IH"cM%]#ERg"%HAY6o@Jz0{gL4X[-Xh4Y8-|n}m>Zk"pAL:ӵh<cw`<]B|V{8ӎ9CG(FU{:}D(;ܮٹ0ش'RG7r#c(+)8rA^Q vm\K@SK)>sq"wx& :d6؆W=\o!c|>jF~H|No]6KB|7ߤNN ٳ(B{Jf]HX@^zO)UF\#c}KVi3(s7DW,b_ϔBw:,{"||s}﫢\IM>}Kz6j~yX A K܈ fTuLsmz$<#% t{@~8ϙ0ыn)[FnM/-;L8QO‚)>S*EK+^*V6y%JTi q0fV9gU? a5'GWR1HGD*\ i-3))lNPZHv> Zl&yLmA>sBy=DE, ñg͜FTi{9 vFol*Zv٧tP*#/1Wκ^vy/Kx#.)򺼴𴾾U7a\oo]QwE5~Ui5j(S"hL&Ҕpe}+!g.R盋_~&K틡/zs[qB 76#"oc57r/o>\\/ӏʒHk$КƋ9EwCzj\U!f!hVFRL"F UPZ!Sޫ( PݘBP=} ʤH%xKڿ/i?!wݗVS^ݡqۧ W/?nٺC=v, Fgl}L$]hٺsHnq*KV)VLɅ$,Ыwnyۉ7X_vv‘S羧=l$O':uFh[C2[nd Gʫ9X͊%0+H#V~ǝ+D)0b'y6J>JոN>kv o)q5R/{×Z%+a+Z9oxέ:o {Uثg3$6AϕU dqaz*um!7 GNcDgƃ 1my9'坜u^T#G-x8c :OqA`&Nv9FUZx yLj2 `b7Fǫ<M0B'儰"&)AdH蓕~'LVVcUrr*iI0DZi39TR)s +& bVGM=-#Q"-l>`FiCNlM4a[/D]|Sed~ )+1DxGK>m1Qo[LT=|k3q&z"gɺ;n;ld/9KxϞfovcin'zW=x f>?hRI8;Xz,=*`iRw@6zHln++ʨs pB&%9x`밐zJ(娷^ê{J:(RY;(f6>\pɒt>k]yU=UYWu'e&J.Rv] P'r'l{?ڞydӿ=J6w$,><=;HH_Uշ]ۮ%Ov"呿Lj]kIrr ?~dR)i1XJK9L+'7>dQw(7jQژZj0wC3`_މ#) ch)Oa]}ž"߯Sk[Y8л}x_$g@fp I+p!V?߿eVN@0k" BLwZ4U]DDmEjaD^ݚfJ'lOzQW>gf߮nV߆kQmM}{t}r+ogj7)/2{w 9[+icӏo`|=Rwiy8dX:Ɓ bm]$u8c$ ܃aP64WbjC}hQRM(~@ 㘥ڐaպ;[䰏aczkJ_|j.JJ;V4b+?GXT/R-)=r+JyVTsdGmʬTLVBmő^mi ջƇw/ORS[*`c,w PVmC,ʫX]w'r{{.obXjV`C@Dଖ[MWbmՒu*칁+Tzq;TTv g a[O>l&ZZq5H+V-iJNe%4W4OvN!-_?ŀBmTJك+ƽڅRM2`= |} @90h'h*xopchVnC*!cJqA?/B!h9w?jqU]+@%6E{pk z!٩r$Z/s*^tT 6 {4V33ZEpP%啣"r4Z%@'Z"yj}Bf$vR(p:-< J}Ťf=PYxh?ȏCu GkXMiE (k4' IiG ɡ6SkiPK%RӼ0EUtϹ|x,\vf¢Ґo*h5 a6oKoF0ops$}N% y*QX{6z$י/yC E]5cA`Ho" `E^5/yC Q[[r'KE9 TpuKǼ)dLPUprm Hwv:]*(kn<ؒ+c(VB&U}ErMgkǵ_ kٍ-2Ӝy#0> R-M8@z`l |)X4ڀZ~ͤ1 O1ʗfc64)5F8ʱ-Rmbgi%o҆0t RKWnLDJ^Ѧי/yCPx*XS#S#x6ekb0._NT@y .hJX> ~3?$jif<rZ57\0ΞwfoO^;d ;gnguBҝs19r9Y\90ο\/3U х kT/R-<5D 2+E61 Tg`6 +Ĝ͗ {9ڕ{_`,L򲁸9:F|WQnr=(Toce_ {Y^p j2^b:4aEl[VFR_?;Wkļ}}0|~b{t=@>aLÆ`JK0RIIrR7u6>O!1MyY|v9pmM܍a=9zȈ󝙏ؐSae|O/n`T ;8pݍlf$1pI~ S#4f"胁=J2$P9b=̬\XGw["#qǫ&^\(v#>Sc'oN)Zc3\R:B:;%DۋCFz|ۡH=,v9 nJ^W8}TR+ܑN @Oq$S[R)9ڳrAtZ}ZCL ;=+BoILK޷6`A$)<2!'rK 1B$mBjtcUHۇJ ,oץ1]7[rju'%"g3 &iC cKu~3z>]He>L+_ Fp\=D3uC6G7> 1*ln|V24eAX8DցwxybcSG۾6DA3aYW>|%vk"!bļC`-(MVC[EYzh(zӯUa!-j1  ͆1T )؆T9n!ւ3vzir3{'1( pk|j`R=^JzfuvT1?o|g8q6mz?LsBwg}s>Z{{*ۇ/;(+߼[Ξ~0ύMrYoꞅysk4Ѱ3EoovdAo4' .}GJެ=T`m JLg7x]Sqgm<(Awa!߸)>\nD6V%3C|wxBVBqئ yUa zR+L<,2I:$ЦeƠ:XQEpmNhنN !M<_V{bd)XNR#JGRYr`2Z 3jmOB$*nnџb{KbO.hF,.v^D52 }@Γ E^P ς!AKb}5rA7*-cIkTiY*6(PVI*lN5J†6,1kZF+V[ytHb9)ClU>edܜj 6Nyk 3%oX 2#ޱI #mX=Au׍3rw{7f-8 Zch%߂PI77B܉ޭb"duS>ԍN0DN]kux?n Gmp ɔuJ5/c(4+} Z+(ܢ9|{yEX$ 8cA=77=~Ӈ1 yx={&{-ځ&mxii#Pl@%٘C#[>Sš } ڨ*odBA1ڷ\)x7`pj3/R8F; h B7-FT8n\|j0oQծ\MڏHW̸&δ){,xCQ`G#2hm9XRf o Z$0_̃fZ(CNJp42Oי/yC 12 6Dſ6"EC|BaEqS}H5Bg " ;.ʬ S3]/1rmY6ŗϋ4ٷ' ɓ&6!~wsLY!4Lu)Dh](_P ӐwC;Һqo5AzB uJMxMqI'k71 D||EA{Ge᾵dJ胰ר>_Ba8*gVQ|~G 9B4ΜRbF @G7b(B@fgaQ}H5>GʭԢU}gCHI1tz~K?t=8bwU&m.|dFv;faY%й/y}gߋR0B-xjy1J){PJ  pdF\<`¤U? 5(yą(ZM<0@9Leo<3Ȯ; uqSS|;Udv6V Y-2#W;=uuusraCuv|)n@t[ꬅdEeSa rt,'Yn4ޗKi38oRpF2Vq~, " Z`e҆1bDƿ1*`yr,!*ő Ai"UqR#sbTbT&:e-c3w ڮ-(q d;׳7.#ș4$|t7A{x?S> 0,z$bL3е/zf r>6fLU,8Dͥ[Kjqg X\Jx.ٹQbWq47sUXͤZ"tmbjT0z#PȑisŮbk N&Pޑ}Jh "chh({ T#UҐ3֤;\6嘠x~hGz\I޳UP fF$TT`=z N 5_ ~&Ԝu4G-AkԌZwp"q*cڄ\ěp jy{['WK嫼mÅZDPo)˙RN 9<5Ũ5` 1/skS)J/H%Ot%m;XL*Z(נ SψUWp8m1񳊩?h~Jj$=ז i-|BsEYt6Iz)9Kͧ{c2OѰ |A1s^+7J}Z{5/ƦUPCKT$ ))Jwc{Dq⦴I{JT"2$'Qbu}-7~qh%_ZE~t^-Ӽ:eodWUa0uy/_oʓ?ȼgc&9ӕVwm1eA:y~.N_?}[ O-z4U (u3Lёv^>;9`cHr/^^>>1;OOtAhFUc_M%4:)hVr20H}H1m(@&e"\&Wwo Fę#')Ƈq#[ m7~I7K1yriu51v/O &>.σC`OSǟЩa^F5O剦jഽّfG`Ke"^jȺDt{:]k0kMN6L3t,5ӑ'ھm% rdve)NK]/U)VBJXIvþ+S9?4Q#m8<.9\J}G| ްO|t!z9^(w,y =wjI}&DmYSHX-73K.hw9ZXDbMr__cޏ0#.Gw*L;>?z*D,B^?}_Xf<=\Zc*5mj/Sm֩VM.OU˩k4#0L+$5GK-(Png2?Rjjd\t"sa{ໝnyO_V>f㩵*{-RcLN8>vs;{PFɢ.,l gr^iD蕤$DBsR#'@g+#]-*,a[1 B̫*S1JȨ$5҃yBp$)Ly$)XdCɢJ*j*~tUJԥLþnj5兏hT>')q'f 8){Zz@V=fe讚_M X``T ELQ,\)..ٶh4\EhqQi"E\godyEa:X\!E>c'RӷA#3gEJ3E޴Z-o M@}o0uoz~!yC?>ho4#oguOtxI/l BP ^b6?nT-cQPFCҀtS΢b:Z\Ȃ6i>pZH[S#dԌopw1Gwa:T!Uv: ojuV?^fѫ> ,Vs$2AQcLԿ>Gp5`LX'%MFpFCDҹTJ)⣎ȤI֭!\!>E8s]nžxL.P:DeP_}J}H}i`jn}L-$0%PI@k@+E Ayi@)si"caA t"3ߍu$ŨS!ӹU:v#AFG2,2AO2ʝ 67i"R9MBTm%(D!sJsP hr>0JԆB2ǩU {>>In=In>.\K[DG6[w_?=;w1R r.=rٹ ;k&óWu}+`mg5$%dve2(-ec3ӆ{GH/608vDujT*iY?g`X.#HvhspR ̰ ^V%zCi{;Љ ";mmy>nt#0hݻW>qy4ޝͿB\y/oa=" >6 (n'1Cx`S~y[hnFfBtl]xqonpؖOqo_js~Pa3d)tX$!?)cn=6%3tv8+#Sk m EtLюX-mQnCiP":CרnEީ5Z /(ncH.2U1jdE!ήcѧHaD^>]\}LI/72>. _>], 6I;@O^  ,'[?U5'(>d}z<8΍Q7@ˆNBj5Pr|e/0Di(LVA׿?j` V8yU64^F۫9${I*yL'NQ&X㴧5Sh7ߝ;_, gk@^+7I#_p@U p_6ow_ftdGY9#=92-ib@IVwU=dHS%VFoAys7C@԰Xuik/:|P> .;NRĶIXV3DB3a 5>>PTZP*<~wqL]m _/7m~Sy v si]blK?rqЇ^xXQs_&lʎҷDCHެ+*M=>;,b*<*Hl/{ ~ۄTNl[o*BLr\[+"iT#;kL1% fIMz;C£J%':m6Q6)t)>*盝߿} [?3ɧ6n?^*y×@wd'0Ⱦc.?*7zpT"yB;|Є/5["yƞ z̝K1tП xb 5(4sBN0́o|v{wV%Bn;MZ9c7'E]%I[מa'q L3A*YMeN(3SK3%Y2ƆI )GԝJךc.9!FIċ8uTJ0w4Sf}W./ qYJ`QOPg$QEV3u(&f6ҰQd9RcWԮQȨϷ;}HY$W׮NHCf۬EjF9a}lPޫسR((J$ϭmBR0&_i#ghg|-!0ygIPhO:CÉɑZdIN2I")Nhwm HS$ yZ&hs1D3@X+ҝvh#/Ҵ\hE6q0w{ݦtH~D E*(pYյk\@ƀ]xˈg[ڀCFē/"r7iHC du ǣHZZ({Z(45yBj ]ZX FQJ`4:9]lԲPz'#H&ztJ=xhvdČF[4 : (>]j).4@< \r{i Z41vJm".5i-<_ɁR_Jx%\8Jy]LS,i?h.K>٪VP\oEj>tU Ls-<@J  Z;I _T_FNcIiV1L 3lv26166΍Q6iڲNg*?¬ Zf,Xl+;Au@Zr5D'ң؛5d8(&V *{\kUHu6ͯힹCUZX-e*6j1-_G#sZ9ybnUD,:OܤQ,TtVWy/PV&x,ї]yѪ3{iDF $F7(ـmc˽yKE+k^Xvs͒/N.FL%]*¬`hf￸wڠ#ZL\.8 I3%.V [H=Ʃj]fNjF(5qKLp2>Vh2SY%k s)Tkw?V&uXx,*'CwA S ێu@FkVtZՂ־  ܈Qi4aFwv]~eH eț9CkJʒғ0ČZ)v܌BKP,?ExfBnj.HCW }5y[Ar"~> hc<ڸ{_JDZw=w!u l&"E\I}+5h\^6 +"Cs)Z/ ͥEMaօPg3i:m3gr2++}^_iC2Xj7D95~{rgs-ުGmt_yX!:<\@X981Sc 6@N‚5[!ZVn NJ$PJ1^ɔޮiR/1ٶΊÆ;o`[(6:WqքC˒OMϗ&K"Xt-4?zoP^H_7=[Z,:Ka碑]Y!3he2X)4XZaZ=R[R ?+WQn`="ݦZk[BR@';aFi3,u\̀"e= m17mSv@^VJ:368tK~!mNr:48rFCkXT#pm#M8Y\ʔqԳqGc(6vfJ&w)eY˒0/{>Kˈ|5:{NC'$V'q6ラJPn+'X`sxE 3>-qCUDk(*egRa{ڠM"S!:!' ΨVx[J: "ہ|H VR'iFʮk%I m\6& ֋ Kn,с$#t2]c ~@\[/p44C6c(JhkhX*9? ~C4e~נM]R)V,;˩weC F#:M Ls'd5\7J؟9O򳬚3ܗH&~)nP` (4FRyc dWXDR^!rHLְ~ *XhV̲#f/q$I?ޕX#ku[*+l0u%㸄&B\[rFĒT3>M"h,9 + {oQҨmTƉZs!bLfL}74e"jXdZG^k!R S̼A )]P2$R B6Z' B%2 $Ep:H?+ |+anP5&RvQ H[ 6TRLJZ˴P8!V:M?_㗶\ln5!y >r2C V߮YN,/owL]A ejFfeyl6M^y9ӻw~]w^~yE"`(WM}huDgt ao}wW|  BI>8;H63uۙcJZ4Ihlzd; RC9<(@a]"wκ@ka6Nߖgze(&nbǤ@1RF(5N;:!|A;u>_F{m>ƥNw|*-0oӤV"5:)pf20sΫٻJ|oSi48Xtrz%Dڑ򿯯.=?cWO2Xy+$˜\2HU=M$9uF d9 ['mv>?}mnR}a è'?$/y R_nG] 'X-KX'jT $2J?$~ tm"~ :$@ 见}(Ʀ[=QF;5ǿD{n Jٛ;{$FC8-ǾFQ6R.\ꞻ뷆8 ##"VXj` :>563L_~CO\&%{9_Ey[θ~7|>74Kq-OAD׊9N&!d{^7:8$73#~A}]ǍH_ N2_lm8_6_60=ݼE̮/?%ZnjZ*)zM?Wu}:!t5lC*T]c 0Fb6Sěc 1/?\8y$x*Nd!ѱ<Յl,\JL~mǐԘN$=Z˜N {b C fxb@X1Y d:,)ͥNJ5 &!ɀ{ܱ/xx& Pn6Ps107m8l=`A}ǧr mnҋCtp\/W2YA|k_yBM%Y&bZ$(+,¢f[(IaF*R Ȳx>\пMmmCc=}ѻ%T~XNbLfUnn8/_~+ĘR:v:%̌lp8sG\..8a"Z$$ *V"4*Inv c-RƖ@P*"C;)VR;|KBcv>[Շ;Zs%6bm-&{Kk)SF΂$J \{Z,4hS'ܙӺ(eǤ/5㊍-/3nVLP286a鄒YVƕDC! @j.RE2gR9FtYVR(Iٍjwwf) r) 3TDn FXBVD ,2?F(556OmhwqdK*h_vDd+lnEjH4zl"}ilƛS9/0OPHhu5Ec|jE$O O^.C) t yB9F's.{uO KJÒua*<}OQhzM[88UAZWR>D |2=,8qD[gw>+cT󗇕^|5P;Cwf?sJ뇫T&G9~0_|p2ܖj0YhSǿxI-C|q*K@N>, ;9)X4\~ۻ%av4%s`IwhIzǪQ X[2˝g$ 7*Nj?\k_[j"H߳Qf^~IC 11m[Q?1>,rJZQ==*M߮bJ*>Fm_'!Iqvܱ݊1aF)DNu'Q"ɦXuܾONnJ.34q^蟲W/~ Dzk s1DҠ',b*I᭜TBԓف{)S 1&0E]̊LSBOOaҹSSC^K* 뺐 D1UkWoզp'1xcK91+HLLJ¢(0ɋ.I K|I$a9ѺjU5*鉿Ki5=ug4$x氘kޤ[ړLAx cc )V-9)H]df9y<.gB ×3*GѦ8d4tK~Ցzu`dGF߬-oG1?6&\ίn 4HvlwDֈ}-Z.jM8cT&8X;} S~Hw`Y2qsN.!t+CktyS^b~b~]$ϴ~ rjV۰kC0[;w&p XU4]k]WXGҲ)/4;u:ڂ'7O?xgW><VO4^"E>E$SO`KR4/O'6 !OA ]Nxx"~T*=_;^)'4#^0d/,: \&w^䥃 @8>5A<QǐlXIΆm_A#3+QqS"ncd xC!ΎC 0N$c.w=A29KHNgSP*4pfTp6U&6{uXFP+*3ȥjD gpΉ@ 28SW:Qt['ugSP'Ij*VU¤D$גЂ>TR@M3IчضM`K6ۦq {-D%fG1 IC| py 36u(BHA=K!(7Ljzݣ?Ӝ=ZuUo~!~mt5;83/O ,b6k_/w}zslL\J2FoAr2)j<ST9I]Lqzcp3+"wXeWVAX:"*8c 8?CF NfܧpN$)?,ruCW(£; c~ׇ8b* FHrڅrz$sS7u7U9T$@bUK{힏2CF'-Nve"{ & JH>"l035_=q|Fd&ˑk GYG8PXӢ#{bxTqe2m&I&NMq,c +QFtyC,kI{$ G#U+FK?'s鵯@#rl l'1}z4Nǃ^# IUܫr8A 2K|$҅qO Ά~\SI#θ6@ dP$UB87Sba;N?$fH6d8-1bb Lz!"91$3CC>#󬍛T sP:0$0Al=b.beFD3H B.pFTLT.2gU$ųt`rU,+Ys&X!RE!sExN˨ΐ2>U-kSU̬k͎IXpeppGQ$'3df (1 Lb J6 b}'UX)t`p{2ZjvŒbo9e ݏj5WY.t5hZ={NakimmC)P",\/O~mZo) 2BA *>8J+^fEŦ~i۴X]`C;ŵ w¬wWv7nUy멮n?kDqW_ʬ+zb^3?{51 \3-|ZXV6|z[y=Οa'g+]EDtҳ1etYqn;=URt3z%'Y qʙgjT3e$zws]Bkˋ76].M /uؾXcnl-̤ 5:߳87ZwWD Ϳ=ՊzGq_fXaf/Gz?tS1iomFUcFS56)=W)mUn*P {N8bpЅN*N0t LF?Ňm_lq]%8A\lEwdקh&-5@ETߧ= lAHxnܰIOlVDB*3zxfq_[xDM'[nUea\=\\s6[y_ud` v{g^ed|ŴkV/N۞i z0qtrwL ꭛y kY vB^ؔ"uwôA=0wՁw.6(qb{73h-r)%8{7|ub:ELb{7;-rݳ)SJ=;ea_~9}֙k6Ė:V_߿y-^Wͻw/!){36.$ڝv7VK ȗCr lfla ݁: ѼQ_ n4Y%=*2zXձCPrM)qŧ{\KxoM;=${W;Yz{w>YLfs&ˮ'Kb9YQ#8\j巽C"1 :yQOےVl{Iq IΣKW>&LYHKvR\Iއ5f/,k-rNemFf,p}'^9Z_ e zհm*uDͳNm5~W({UcAܿP=y8_侽q{ j+xq&5.65Qϫke\~@Ix>@4oأXˍYS퉧cO-ۛ@CxlDCAlH3H$j& IcU|pcyn)9d\e9m5 3hTwoMd(gSfh~_ƍ8J珇M~Z d?-Hx`~כ?zgp۴t;7o Eu5]^Ɣؔm] $Pox]t_*{6r!/Òp͟d'D`uvmaCF9+4]k]G:mR ^صY1-,,Y*3.sCg;m\;K$华]p[^ 'Tp'<~=(*to4a߶f;^ :tnckf+a#AF詭NcGY,F+bu@y7a >:>4p 0@%׳`PI"ͣq "n0qv&n7>CL**S2?{8\jW&㽺%;2[.ǺȏHM@J2EK(>DٜY4_@蓻٦] Sx}!c~)e<kvϫW8DlidŒ&ԸqKzLk|:uQWo0m%ؤ1"@ki5(Hjg`iEHI*L~ r~]ZtYu62kêƕ+Mmmt෩]7v&~v-n2V|ծh֋T6 QWO&%b/P5&ϢE$KD~},uD81kmb4!6Y$-0_!J k%(8Kd8X&N8aƻۧl.r,cۧg]+$R=$NلKsDd hhΞO4l wOpD!ڧoD%cﮗ $Y6Yd$ "~O'@u:`LΫݭW~Asvw+yڰv?5!4)z2CؿLFFϺvE8wşқ7w^R'/%(.jZaDV[x 1WP4[ftok0#w&$5gb̀v%hbe%Kz)q0ce KOr|؟oy`ElcA#jLt^E*+?AWLP%e 7i.猶.% amw:AgY.HCJ ]5MVWVElU5+CK8z'Tt&6SB,#u[ *D(4R])̓'S :ګER mPEt;Bo ru󣉃z\19呺UIxkNm2RyԊt"q9B븵o42?U!%gWcrX.YۊjY#!̷rP}U5s+:sn I:<9n}gJ(n;1QA*VL/Q$Ÿ W(u$Qz#T&(ڒL 24Didv%UZikQ*)K<:ggnwB[gX9L!9N)5Qex{~̈wje#:s׆0Co7*cT?w*ZMID,Ϸ*nPob{߿Tq@g7&Pur;w`羕ũ9n/1FpfTGG;7_5X%ݘE䍠JqK wD^8`ϱE -7jÝk@C85 ļsKf!䴣6Í[)":SvҢ T7\й2ܒ5 ak+`DP#۔"T6..M1hdY=aԊ.[B411ݪ.NBu .PGp[h6n7T@> ~6 %xTd+#o\zK=4X"@2$ABXvk]B I H8Tq0G .^)(UO2~A*ο/ER !Ti0A#%yr : o^ûP}D(FvCϕin."HxvcpT`1 = ̥PlDP })x>GpE|7'9ػ+aAVsk2R >^@֨j[LrjkT V+D?r'<^1{zj o5PLsoy-uCQx5a~?Z|"3g6ROa}ms ߽ySn,4[X] `5؆w r>(,#}8TSF/;UeFn!nUBRȵ[:O5`v,յ}&@e*} 8)%dc{!/Q}UZ`;d>nv:alS=qIR?{or%)ebў1(qN@8]NݢSxo ֪ ˤ*Jʰ&sZ^*HY(hA:XAnw (ȓY$lQ[#MF SN+H-D':U͜&MuZNbfA:j,H5}<׶1qY{ 2c,%3VD;JTTZ0B 2֫)߆ Itt-/KzQ0 m1G:];L2)eWj""e$i;}fgJm~+WX#Y'DbM-Mm&4 =pw=uAWwa$56w!kCyk+,WXJDmqxN?7lHV='~z\y>grVɊT׸O.COFg_wQ?M?Y-(贄V"Own2^BtÕF圂<I…JY/`&mf ̍ />c<y '1G.ٱ.k{W{yu(\[;7[h)1)7:w7+@swo6뚫@z.iTf ;2E6ufM\z$kԠ"NzZXRt+!o'&֯vdAx㩖~ޫ2~)<,a(ߎVڝILG(g~_Vqf! K!Puya~Pw0r{ُ Xf[ʮmX!) >O7壷mv`oCleS֐q#S'G?nڭ.bDU .?biAڭ y"ZKL T,?v˃ѩ;FvS2mjEڭ y"Z))SbB_ƁGP#~vv{|rEPz?̜:at?5Wêbރ^޼xMw߅[(ѻإP֏'%!nfTs%zK8)%*dH/%TS.aҳR2X&RBJ_L5 (<("J RigT3 -;>W9B!nV[ȍ>#VPdbk"fay-ݥ}j0_PY8zoSJm~dG?|'b6@<C!q٠McH; $Oa$8SC?P S$n-O[dɦ%T'ρN݂ol%zH^]|~Z+&fc(Rb33:l/ Ic1؂xy0ggP ? grEΎdsHJOa7CRVzn#X_ e*>Y_PaN1a C[Ƴ<+"'=KT" MbnbUkO!$*G0b io&( ?$O &FEY;ɪ!+5u{ JV]y1wM=eJ- nMc1k#d'LRy*9œoU,mj5#ɯ]:wC;stWKl>!}\}W"DI&Bꎕ@:Ydũ"I 8QJF\KL3ʦQ_ҨͲH I:6L 6B` I Ca)AR$h-hdD/r, d~dg%]W ?++ 2lyB{(Uz%)bpCBٍxdR]"&& 7_NS Ir3{fV uy;%AK'x.,O47X!1 3f~=%WxlHJq=clX8njs޼x$뵏) ]U ׹Q=c?)7[\&"Y\MozM&e*6 Q_܀ozHwBF^NinA/UÄ!=~\?ԇe^B!捱P~$Hy9 _B.dy whwf1Y̦q8$3 A4h^Es3ޏ0+iSREI$F$3fhN3SqrF~ Ap~)chv]VpBvy X Y޸rq=xI:T4gyFpU k9BbI; Q2u\1Ve+kn#Gݙp%Gllt 4NE:Hl!v=IRk-mXf:>V +rVó}<[E߈?%tKؤ/}cp7 e.۬+B?S/.m|ѤS^w~ͩD:;А^:Eڲn(>X7_ \TM[w^Lé;ѶuJhBh WQ/nm0EcO Zz>Q\aS>.4䅫:'umeC{*ѵBN~~kK򂣼0vu6ryKɑlOlɱ̈́c)/#/vOq/u9@X ,k5q!h(@R8b<؏'C4=>M b%Z?9{P}N;V,nX˩%CJO4T kGhp^!T"/&Ip$r[VqNG$jĖ",bL*::ӀQcNɉbK19M@ uij} j)wĘ;VA8fLΩ{P+mI0ks<M3P9H :]!fa0o_rp#0 ?t9&* wHo&y_3'K4D^ҲK.YWC>ڒZpCP/.hT9XD *<9А^:+ߖucZLw)]hN]65%p)gI&>Ld#9$FG(➣.h+'r{.i5\XI#_%V#NK'ȡ$%0$]Ut~rZ^bjp<\-L*%O/ew{H*Azz}246 !d}{.:]'vQ)AM 2Acb7D{,>^G @`R]F4,:n7 U%jC`(Oy( h%߂W"Mq?^@⭢á< Vm☯[EIv&0'fwk(I*3H$/,*$P4*P優Te"EE*EEU,?6 ##ŀzHd'> u?SiLޣ,m4*K5'kx~z@-ZuijH~s=-F:I?-0?,҉jfQiߓ#}|fdWugCm1xRk|>ts]h;(*$%'f|a"s)iZf.GjӼȗmDaw=kQ{qe`f5Jv5wvKkPXs'Dh 2S$ 3jyҕ WV+YvHJ=>z=np0&Kphys>P1"ZA}Z$ x-kcfEj8ˋ"]IzS'ؼ We;/Ǔ[9yIL"?B?bV*,櫗L6ޥmէd;S>bUNқ8;lUߦ~]){|[m}|CcLcoL?ƽ:X>UxY*rLW~ 'yHd3OJtJ B4h'60pqG?>- wNp#'J?1M44zyLC/iq4( ӏ;9pˉya K1DS/cXφCZ%ΪMZʖh=/]24:XE<{qx݆aRG;3ֱtwy_|NIEd:ϾTtOISwN7>ѩ'40G8Ae`Χ831 F$b)L?6J&B1_j]7J;.ϫL+4HvҌi8$+狲Ҽ.ʨQ+WƷIJB( Q,Y`EDV8/qDU9χWWzZ4]K>`!=֨V}Tާ#byhz ;w鱊cJ/eș"UӲJIMqH"gTJ"nUh*G=GP}TmJ3Dn?LחuC&]Vx,[ﮮwD=j>_F!*1y|DFo? h0ofeZS} " ).̂aN{dG ̋ \soϜ6*D2Ry0V:yj˶b[p41"tЙP(Xhz.7 (r y.=C`A!ShQxFxOVo/Vn8̹1R *Hq^2Tb}Je %²`UJH~1(on5򂀫 !_An`1Xh=:QL=h}xvfr+b!7aC|~W୾8|o_*Dvv? nMʽ$ise͕6Wf\NBա.^ ~)s=qMݍ0['4|ˆ8mXKԁ$sC :/u,FmJ&e˯` +Ua}kA~~kA `(ڑ|\ї,qo&ej2Z旅s`.cY%<5 *x IùWtv9gG_\.]pzb6xlMңK>8,hSjJ09'<8E -UvkԣMTi)SRܴ8=kRp| N7yu@ܴt`ip*2di4eit[8s="l)RsD~U$zԣM=qZ CBEH Û KS#ͫ`Oi.GA]Q>芅KE[c_qCPqGoO  [t:yH˫WyHnUŮyHK1!5dx@ـV~/܃6#X) Ob>D(2Olr*٤wddcJiCf0fA2>u*Mel >^GIg}$i1J[陣U$Wttx ֫u)+H^>Z8g%Fʫ H_ZA}0լz/M f82Ryu]^/;P˛?ڸO7FpNtАt6,0fu/u?aU]VlyvHK= @X!U>N&cҎzAWaqK]悆b\ΏGRgyJz$ORdRjeLǭC.pMfZ!! ,ypYEuW$ ,8Il*,VUy.^{=Ӭ *2EO%Iciđ< tK !"%4\´ 95e,U2?@Q{a>Lꞥ{?38mM=PeTݩGޞfM|Z I/-.aW#q27[ J͵$U"]nw I+ҭ#н~ zP7Z{|V $wAtd`&&iT!U ^Oݨ wU+x kY![qչh@0&ԊόR=jaD2$%7V~QO\{ ܧGel2N}_'>$e%N~pc;N5=;}:ʋ1/6yYȲB@͓ғO I?sF?X:iW oFQ\L&4vjrTw?PK'}l+ SĎ0?@l8۩gQ,C}gwD0'u1<[/<PK9O;'!Qޅ,Wc'Ɨ Gw.W #\b'Ϧzu:?OOΑf'O0R|#&{0C[EFDC G(AH(oQq\*T({8 |%Z'wU VXPda"Sֹ}hχ9#PȢaASAS%!(P9٠Rjd5IJLcnHF,@2 XP(a( 8b,K#a< {†DaS6 3"N90gZ+Pb!5 -h@(Xi vbPk4S'20IV\lNS&iF4g-q2X޽z={ikˠd_뗳ߛY`fxݾ+]5 "Ho4>E)CLzf=.Q3x(_M0D0*^*σQ/8c98Di$Fp `rZ8G$5ڠ7w.L]n{^d)%μ/_]&G4ǹnS. YN` c:Xlfگ`f݃ 'n= sd󤒉.dW4v./+y?dݼ"[RgNUy]* ߊ_2W@.11B@ٳf;/|n,h1-7QM˃y5Tlʃ{TrܒL* +m ysq=TH.Q%SG6Vҋ+U?"jiWLMHRAA#Yܾ BHʩS}ٹީS5DAPM'aJCXUhXeoM5Úbrr ХoxnŞⶤNO|f!zѳTd&Ӷӆ g0u쯚wAr놗W_:fWD˯rU@UaԀߒjFD rNV?Q(Ǩ~@ԋ _fb)w&Ϲ8C4ܖ %0a]$j&#> gxp[VC3f8Eyi.\]t+cYu27w$Ph]SН8OJb邧U{[=Mx81rOfS滁)?Ll;.2w$!?V)I~&y4C[U |D;hZBவ[ڭ ELPIuڭ*>Se!WZZU/aꐐ\DRZ=Nnl!GO">t?=worױۯ5_5?Oo5qF@嶌^#Jz$+I< " ]2&!> nA:z ,22b7x ԅŋaȎ:M0 t. 8]zpW?F]ڙ}zk?Ouuox+׳/zA^:]_,5(|VaYH ׃}'zZ97TҎͥ,X$~J&w#@<?5Y"sGo8e^gkyq@ ] }ոud@?7Di>)^;iƂ'Խ.ZV#Ǚ?{ -\4 ;p1q/uMm2/5=/1'k4X ~r -̉Ӡr5;3DQ:TWp$q շťWؚ{˿I;4;'>w7YUo&upTuYl/+FA8=(_Bݿ/$;Ќ76ʻ;|}=ry&P~py w7iSٯ{Srğed8^=+|%-]69AP-R:os[e^ M9/Fϓ0Yn ̜k>߿'0z~77ܧwm-~\+|q7h!v`>׭U/@JjV=-8&"Hm._5Wo Nlհْy5,3ͽ^1[%\s yy5(n|{^C=P+#y䢘#礒HHJE 2.wr9 05UH@ bH|DLwW\׮e0OYjq x:/ ЍEHc}B;[50#wxil}iL{x4_5a0ZoQF(e*nkͨ"SOR7d#(V8* Pu0#Ofb 5gBkCXHƭ PAqiZR6<YH}Sb4W>XSGË]&Yr*35<> 6_d,_tћs'EJyUH<; eY>/_ɢ*i1VX nDZqX(NbCH* qwֱ;nWScoI]jsDe3wHL&c)@hړZW5wKuֆ֕@pYg8zB p 0-`lER)%mB9q!T "f 1ֈ"Jlwx) (@"H)͹BF!(1'Ƀc!H6#Q7{bJ!Fl}B8#Z MTj Q j+-X %ǁ0Da~) CE5aEڒ+F1&?B$2tW C*r%SKG< 25rQ1}yU緪+ 2UݮVusK[JL t_d0= z"W>`GY`t ەdH])ەg޷,=΋e^`8+I:S8Ύ>JgrlzAv?=),_?6 |[=wCV!PۧwK$-T^@5h9UZaL>i͘2VP(m)/|6kڊ>t9^RejM&8h>@z퀘%b iT!w*$sqKXVbGo>_ݛA?\|N֗.:kus,O)%u~ Kyt m.)]h ObE1 U8/}i5EAhe\: 81;cc,Db,d\yYNkkWSP~ '|B%ZS#X0! =S)t,ҡHIBVVPXܻn4?i['q{n7mi<3 z[<ˡ 󜼙B)WakD| 4TYlpD( b#Ic@&Tbc? 9n(ZJVF[exHIJ0CŨ2V88d"+9`R)4c8rp3%I Ljſy.%BP8G S*R"a<ڌaH@)|NIqυ,w+q8&aBJu!MY:Ejs9E($ {La r9Cyww 1㙗hv8Bj;hks皈6x't>ɠ;/J|ju ݵ͝ؾLP,;\sV8Y !mqKI2-XHa#03fF ;>oxry";e<{?`G}ŋEk\hx?'؎3{:htFFvv'S^~^İ&JY/ "]{A%C/Ŋ {=j] ZG$,k0f1H)p̕ 8 N82%e J,Y6AXѤR24}#&"?aL( iA((J2e@S"`?ٻ7$W ffIGz0$z0,d4j5גlڅf.ʺ(Q]*222CUU$=K$mok #C6A* F2A&h"X$4 .P$NL`фRL1K !8J*(8F,È{ON%3\E&KUj<ҒfbԾ֏+MAbS2z)Di]UJQb 누֊5?ʿfr٩v|ҺKL@ t9'B)j&EnD`C[ä\IYfb\nʭK@(@qe9cx8 kGD)NSRFj&֤uWP,Cp?ҩ >`l7a +VU>UZc/9cp9#+\z\N@wI<@-+9`9aTiTRAOP{>=ܟei\lM=QК;< ;D{H+b[lٻ?'R煙-$KülYijW>5~vѓчۇwmKtl ]6.{je,=6֍wƕc~lx P u8V5s%1ֽF()ԅХ ul.%C cI8w|5|yp @IfOiVVgdhm';4KPY[[j6qY *jVzL)k[ǯ~U B `RKM7ݑ 6#v(=ޏVq|n>rތq_l7O./3֬f.A3]f@gPlHvHʧ%CKsY<۸9]r{L;=(rSBqa `ޮwSn:N3xC5μ[8wk!8D0pgٻ)R\BugTn="7̻bzz&C6$zK_X w&ry~=Ξ>wՅ#P)Z|~mz.zvysuAow>Jh!T cG4(lcZ(F;(AyucVspE=OQ2VU}?M *V~Ru:Ta㻫]$_W Wv$v!jTX$bEU/=0<#)BjVk?fU5ŝ~VB! zc}{=.܊h챂E{+bK}U쇌qxjSS*Ǻg Q1bkJpU{l[PX z>V sԳJG3J0Qg.åy2ZMcROR-~Q=yYPж]X?WcݚE7&O,1aea]3ԍ},͋e1wZclV?P~B[{sx>}qX4<%, bY 3[fλѺr&(!1ֽ\7bY?$kY`sy5H&t]M%`W3O|"%4T/|C@^+K*6chMsv@`h_{ jJ, G'{9MY_LM2A/QcddcŴb6"xB1zH=>9cxug^&8tgU^91MđP誣ūU %\\ T[6htv$?K4\d(PqYw(N4HW)I>*DRAFT0+L`)2j$!L45(*?6Hܑ2Hç2RlAmݩsx+Hw KB#Q4IX̪TZřHӘd "Z&v.cFO5&Jk.PmT:%ȨWb <҉"*͔2<㑉:J")uNoѳ뉦TNwuԂ~ةUQ#}Ni'31݆ӎI=ݑZvN; ( ᫙3"iA(,31}?\ `$OT7H^v Zfz}-Cm#|_rY.z>/Wuk׷z&䐐S6}+<ۙ{޼b1a}}og"s}qFoi&&zaPE9eai'WKq쟞O'ڍ$GH>JnʀZg;Ns n+.=׼Մˋ.[*JYw;#\`1oG@ C8#AOd!sISJss2`Px+5ss؎_vz2:DHig=$rQ8!5{\S-;jgnVl^w\ZȊ[gnMWwnb4\6gt֨ZM9bD{{Jp=oۻRQ`ZꜾn] E2!Th[@vZ%HkNk8^Qkh@)l|CKT !kdחZq*jJ\T 5j=ȫ%t 6\ ޗLR_|BQٴF1'_!)˦Lq:N3*yޢ/gSͻnޭ| lwS:Ϩ[9ϸxz&C)Of Tq0s_s07#xrNBڣSǤZ8rs9up~xm(%@qe Jg@)#~(ͥ/(=orRΝAӲ6 J~3u6(T PK=g+5Wb0zZ¦RTbLA0 {K9!xR&PD$ PʄJsQ Q WJA<}i.\Pz(ERteVD)r?R!59T^AέCX݆ЍӲ =.(=G޶P8H'Jzۮ^F)~(n ǥeQ \\0fcROR J(P"0P\]5J{r^ zZ7JJMV'J-bŗ7Jz0ue^I\2QR? S=1.RSLQNeD,5 Ee%H#DGVc2IS%$#1g&5hBTܑ=gYYJJ&5eII`lǐDX@*%ȍ(Ɏ3"e+1S;*V:<Әr_"DK`$.#$UU͐$-z0OOxiCW}WԙX@;%j#-9lIq%ǛdwIt7^`an9% il\^"uKI+W|J˫>C~ЊV3E&˫){!\*LJ?!\hY~-]֝^FoM{Y̗&u H: gso6}zp sSoG5[c6یaq$*}sTJCN^Uߢ"$0r&~~5mV ǯ+BZs*iAh=f`vE_ݹIHWj՜ zMzHcF]]ҬEA9$--}[?c aP5 GK.~)~rFP$³bbްu(:|&lk+ظJ.mU;R4"ˍVS]=6ǯOryݬVG^o-*Ly*qlh)]eReVeu?RF f'mK6O[52X7ր 1|K3U)M!"iBexAQA2x45q,dQBl DȩL`Q3IO\PXㅫtmRC>o vR<޷Y=CB(Շ$C}dfnnٗddb{LR{L"e&tf9%¬Ԁ7ިEݎqh;v/07 0l`4C3vPe(|N+ێT1&wzH7EƉ>ܻn/.ܟ=ٯ|XnB $̾i2P.#Jq*pu!!m?}nv!t;s|?e*XoPYx3~v~ $(gT5Џ/,pq^%\"]eWV&ueu P `UgSi~?6c'Sڣfi$6,1*1fL )&A !*IMb0S5BO<|/՝V_*sTZEv*PMW_G?lfrYVW]XUYFw&m^>aSL>Q\ؙ$ӱT$ ZhLZDfXW L8#;עF2<7RUZ_vAbq0$&H_p}>[!ufSVplx?z]?~ūׯV8$/hqc5$l<)>L$0ݶnp"-5Dq)d%Xc֙8611ʋ^J&Cmt+l<2,sR5ayG=+j:?: ucrɇӲ~'|V32BO&_^¶G5a]p(MZHN.bQh)ДpH8tRM*|ɛ<W%VG Abjb-)eYc"QQV4cY}Fq0Px7E,<*&Js ! " PPYy`2{E+vx ҈B`&@*;3ӵ=$_ћֹwy}BG2W\-x6QO0QQx=koɑE!U wA%0nYkZI~5$- _b{fH j5zvjz$LJoP 53oXOOgA8EF$R>VB $/'W,|v0TR#B_owwV خXg]2EлkQ+iش"GHKH(WvnX-*,ԥ|(whu`4`JZ4*ɷf%x2/j=tw >N1 #DiY둱jN#2 -Cx4𚮚ͭXgIQ; 7ݹ3Ԇ'ŚjMٓ/Z~ؕW,Y3lQLMMn=+z>{OQ 3IEN,dP՛Uf<w 9PUz⧏wsVM)(j?oo+_6GQw ݘ8%]Vm =?T0b[Ǩޛdֶ$'+n_mffUl 出F&| gp!no3/V]T.UըuQ}޺mޏ&ޗ_N{m\/^np*n`x44$?~CFF gO$Xl~qSu}iv(Ҽnשv(TȌw̧% f%cc<[bpj#Ȝm#L^w)?FST[4vmt q2ӘwhS<|}}M( ܈d#dGRHTSc`M.E%^?hGV -" PBKYC@FjNkF֎:,8zd FB/TASѭyF;jͫDCi\yQiV)DFGHQykoU<xKAy*ƖeETFPĵRWDq59+4)+%G"OZIWeKƓ2ڢ@%$:GuG^AcRH֧G θF"+n$R+i F[x:UѱDF(AՆGXkAZDtDfjLBBj$%vw.BtD$F%u1#`17N+A5216H8?:7=oJI?8g؜8c~56'csֶih/iO9 ZDRGCi FL3V{yxyx}xxe D19l˯'cakKQg9^]ʽNu>lP%rt$r4yB*QWW!QW]cn?ssL. 8}H0(ČWhs! DXk:nwhl{ê"sL^ ,m^hwvPVߍN*QYu풜PڛWiͫέB]"EuroBaWd;֢aC볂UVt8dz:Uw,zq/*?LATezB|E#GlobU\"^^%>XR a6asK+ " E ^)c8xoE׀GରiGCaR۱7_cVMCd*QݹXZݪH}6*2(-3!R.QO2w9N-K\aK#MtzӜ8CpN*쎉v Z*߄SLU=VVWRAg93lyЕKkAF!fֽ&ѹǾr,va8L #Σ]*|Ѝ9ʽhTlz PrؔԔ y6lB@cpVV+>/CproIR%u(Bο ZGP?J~9Rk*e\UUE08KWJ/ge% 'ua) 0=zȗ GRX/VupaJ+%SpzV+g=U}c|B=X 3Dq8`D4FMę6{CM0p*iGn]bxY~=8>FG#H:Tdxϓɖ"3An< A=ju =IAByˆR0>Fh:]W|jQ8B+_f+At9e!ٞ=C 5}#VoL'ᗰC}q׀~h@Y09!q?[-O;$G1 TwUP#zM6?6ίmtT>!q+"_D|]I搸 eP%ꅶ9YSҔdAB^{ABJb Yĥn4$.uH- R[3I"p+(~ZBf%T9mr}ԑ @BF1wdP兵yRcC [T`$TZz,Qb. g`BP$_(*hm2?}%f}O_LkQ_lgJ`wY%BYadkt;aa ˓_CHA#\ -˦q]c>yиQ%--@GT l qxG,wwrȞn#Zq[{?>λNvRNMgb!ЍQQv5Ya>Մ9d:Á˫:,yeq\E>-n҅ ߊke ra2'2w qs&ǃX/qt f`:+d_͙5l^PeS3S".˃Y)WҚrTA[:ah1'O?1 U`$ 7) ƹ@UM!,2|+-j7X%Px=zvAUJ%~>󶙯>2>L۲o}?23\{WA*5\^l]@Z83N 1vr-i&$]3X+ok=ff%:'Y<:kʲg60sX7 :A [|P%Iujf7Yy?;^Fm$b@jcG\{tdʼR JFg"`?E*MZ?L&hk Yaf$])2Xۯ#΍Rt$%d1HXj67j9[6uQwmmHyه#^by> Ѝ޷xȲ"ɗίߢdrZV##C|UdNjRƘTeñhcU3k&JŘaTѤ3RAmg-<5"xd* `ЙDA`RvzP|"ߡ\A:D YB&'RUakq2̗ gw]B$0n4< j`m<P3dY$Ӱ ۮ9$VF4R:JN+6X2M%}*XԐ˺vidgUj傖AQJDyUc֔8D0mRAJKb\*BvŹ".(gWYNӯrZgu^bK1.f#;|g+@w[,VS}^V2*WVH-mIs˗bn&;J~.a=fGaw~Nv!=7Wfb'ɡq<[ t5)F^pyo8z\Ӣe&@F`Q((bЮO>[0{ J遝v" bQw֒+FT ,\췚ʐMLO4o]|7|K!On-V$K"ܚJ$)>XX+Pe~Fd?!q}`ұ-1,seG`#Uh'$SN+PQk,etZl;FUj7T1Α%{P*7eJ`F"SXjzƔ]K4l'!WʉY7%`'|6#?-’Eز&m#xR,ϵs>y8[ bC4?nPO{'U0k'U?U[*e# ]t9nq;}4\ŌVW^LbqU+ŇƓyYnwY^];h{/e parLWMKZn^.,پU S\"ο~8['[pw-#~{7A);66]>%'D5B=vp?̈́9wobq)>O׋uҗt:P*k[,oe)PV )3ձ];%$Gs]r,$_j_R+WI9+IM-U@H8O{ IJ8ks"؂ 'ɢCMm"7 2755X>,*BĐ̥ ̭NlE1YI4 )nC;#\f;뼲pVf~xY؞dž- `=whxY!n5co\a֣oY5tUȮcˊݿ|OXMiVr?񰜇[{?2v73wdupF EANӱh `l<*-Myyv,0pCevמ6'd$ELF1x!Y2$I|iGM|,e\b)GudQRaƧPE` tBSCow^ ]j&/_u葋$ZH)EPDc 31\JF?tټ07zP3I;]$bHTQ*SLژTe nwI,{oQ%UVPSO7!("/?=׷q7^^>u.?\(h,t[\:"A*HjR&ZbpzsYKeֻً.ʖǛd]OAeT_RGM)lrN *z_"-w`V6 ^eKoMx2XS~༾?`E)_ ,crߚl,g]9iN/rY?N5O&){sa3Ŋ <ϟf睧BV-) ;&UPlXMyfu>Ĩ툇fg;X Eg(XHJJXv;[Յ'i2[/KX缯zJ¦9:4Y./X-j_<2rw2>}$bi> bjFvRHxa(x+O&x1}o9Ԏ9/]UT- E-0I!h0}$6'dx~=Bv5 _uJb`% ٧$-AUXa)=eN~ǚat|r7ѷN"ZZ3 >h-R =Ef(ZSOGvd|ҋ]uV$('^ [%Yc%5z0'lߜFއ?|Jr^u8,lk. TEDMN *c<&<H|M)*\zZA(Om?6 rb>wD 2Nm14kHSBi`bϳQ~ǻw;Ow%#q>%x":ȖQޏsNjJWr eN9v?cN}$EXdW _@ kw)^B7Eg4No}[;_pRQH#V\ITVȘ@an eϭ{| Pjc ͒c [X9Q{ptLAA\kEA 0l-t&~vgɁ,p ̒dE!E/Xߴ3~v4MͨH\ׄ{#k2 )X\zUoc]h}.D]oO("TdF$߃YvźsΨoj@Ǟ DXc6ɾo7zV|-G }kPC=  jIE?v5>f}WO4ed'F0*@ ,W2,$Y}H<{뒆؂m4_D[C=,H˞~ߍYT\ d0ݦx,nɺr#Erȗ xWDw.]ᮎlwK]VjUl"FWJ7~8bgr3I9 րd%\<;.ytsʚB@5t/:ݸ~ԛ$T>/gv|Z{9 yQ(LdMIɎs"G*v?tϰhI^Β'WПƓI-=\aӢj-A+tO:ZW{ow3ҥzjWM7B`m8Bs~o^ fe_j+ay[UI@Kޕ(yW\oTQugb>uAc;\uXIE葨eQCke & A?wJ*i C%$R@HZTAHzk۵t̳Ź/=j\Cbp~i$@de>%^ZFgһ+b%MqYr?i N#0ENE5ZdBH=پX]>r#$Uqե}#4cs#vAy N3YY}޽uQҒEyWm-|aFfG$q2% {mz, wV$mv/b`5v\>\aM;V0.$$*"qv6d$`Ú|Z@ȌQd#Ȉ^,t,+k@+eD|.U5!G)$L2_Q+O)PzHۜȄJdfvvJ@&̫SowZ+`B+Xf3ɤ@&뽢Ig3,n P231Q$U`"ZXȕk$@~Xb2 AGYEK@R3'!)VI!7~~o>(\'KR BD`馡:JF=}e]H*b,>7tF=t!/;H#IQHzKDSʅK)fxR q1r,V0=i " }Su%(;B1uK0]"n(!|&kg?L!fI݃|+ŠG:З^#D͗qNbz x`Vl:'2I]ݬB/KCV+uٍeSx}jh5M[]2M:UC!pع^cU-ځ Ãx6Mg[it a9`M^x܅Z\oOEn^|wJN<YFgvOE4pԂp$ rqʦ"҆QH>ɊFgrVEc(RI_|( J綩P U[\%H &U!>_}BQN }#nDtW@[ە3ۇ\L!P `#'ĭ`Pno|>F:Gg͗,#( :%fc Z1 j6*H=*X/L(!кdAa{}BjH뢷l2ou\:d;a@G[5֙'[`;Y!y8œ%3VBہyH{ؿQ`a$jrΟTd@'"4>FMaPuS&oU 2ujGf>eϚBP潚j]['2;<ɸVmZ4 f٧oCЪzce_~,X74/}7Czh|Th*s!Ў@W$2υ@7TtㅂRbmR %I,~4n(8;KUW-~{*YJKxѲxa5k)5&8ѳ.Έkݵ/J9ee/?~ksPck@/"Mp# i o 'R?s?MyQAZOpLLQ?O%u?57b^o_WMh'痖,S#)YsTv8=I>}'Ws<˳qݱ 9v89>ȃ NN?}}g_oǹJ)8;e6w Kn.AZk-Yk}g9e}ap*=!?C^|s69.Zԍ9732*ONŹ7jhZ՚3Gإm%g&[YRYaݟ+vckY; 5{AQS罾B9%z=̏I(szJSr ƾ9ܓڒoƎVVNSG[wrQc't"M: $0"CI QђV V !pa#H7uXҙ>Oɯ&:k?PYU=PN" 1:d2Ɉqr-ZWE; ~e<5J]}S=7H")?\غ|~ 7I~+c~ l$#1VZRCӍ]w";imYG;/95܇¹-@p"f ͛(2jhɑ_F53ùkФ:{zO{Q>S2䫗N8U?r_ыɥf\'^s)|qƎMeߩYs(şJ:'Ap ^(%ҲAxvpmi3Jrf(3ۖt CnS]_zWn#YOYOo5mf9LOQ?8ɣyK(V:c{CNz }C{=yW+c IGRF^pY"r "E'!K,z ^?X:ų~u[dhzQo6FJiZж,^='S$) ^kUZk PeEt)}j$P :zn65/ Ch4mmnHj |WH\i5dgdB:-͡_Kamf4f|σ~O,8ýݵOQn@&Tz? H-qL 0'y1{ FR{&TRf#nQ3+񪙵y5CʄN1Z1!F$~)F}Ǭ@?~}쩫Ԙ` 0exGP[%|;=آ[ݮ~ebˤ[pehI lq2oYN~ty'@xg|s){.e;nާt_nz8x=9~7~~?NGqOsp[6ĭ>4ˍ[sq4nU娕.5']δYʧ[Ի&dV:җҽzQBM\6Fb4 ~_':ZpҬ-n(ohCaS F@JFouٻ涍dWXwʵTm֮8N8c`zxSS3-&v(@OAOLiiQiR#D!JI.&\<][Qb/ۮ T8aժnWĐnIY? 0nasUHqL\+bTW%u@UI0kU?8 ~UIvmQz~b,,ڷ*$VkVp~H鰪<*nI@GB6+#{=R Np)-"^j@:~x>TY[XRy2*e  &xͧ꿭 ]\8oѣbXrL5}Vp$YٗܓW+B#ps*VRCj NR 3궏SumwNf~Jp{'juGfStSjI~Y#Į&B:.pjWf/V.ٿr*xWavddS0M61[{h[h[h[h\ ^\')) DY)%LS |~?)/JZaJT2#T~yWuh Kb1AwUY!vj"M43G[(暛ѡpn|mU{}QcC!Rʅ:ï&S8/aɫI&[*^b PƔ̩mb2Xyƞa*JV2|td!]6ҡbuHwF&Cd?1Hm由H}HG#&(Zޙ%fYRQ\ܶK%Mi14WHwF 9>,٣M}fSD_CD|?L\ W_S<c  s ( LDu)1<繡=D /< ^BqySifڈpTǚ@7FJ~?𭍨6kk# EH p2@.ӯ6rsRkEM\Kx0W$EPӒK!x@m<*n) y>#$%ã## 'Q3jH3M9xZQg3=PPBc]}V QEYK2wGބ(՝[qEVtYq/taSfipBw >QޔCVEu6HZ ]B0jZn`ʆѦ8۲ ;vB!5Q}: /cj[XkR"BsP ޗJ-)AnvGT6^_A J&lѿY,j|գ_ΞRU?}F-ЩyqmfՏ $3?MZ&z켰~+yc~&`FcfNh !Qj)u8Sy7G>#Ļ}i ̻7;nm C4 SNGˡFލ;[|*B ڝw@3hcLI6U$zEsmđsjGNrͽ:'^\*;d/.X"|bww] qK(9⳱Iu}}nft};<'QN>7/^`"OˋbH'y~3ZqRNMy;or]&ueӋQ5jxM[^+`þF2#4+u^2IS&pISM .'\"'%)+LmA_>nj1_O70M&z:-=f}Kl/R~@g˶B+FʻofZOC |a1~3vqsNV'U>TۇV|5Mg4MYڳ r/wV4r3-"vI5~aد~kA h]RX{KOO8Q Ϸ2dXTSqB"/QEAKSi4BPT96ykʮvy6~(HZ(`5BZMshzo@:OeKf ҼԴLE!IDI($2JW,/3nWXhƹ#s7 PP'nyWZ'_&4cul>g32#{8eF~IjƽU6Nö]%T_/%^V bT%`QL$Y7SVOA!鮄؊ v 2TK`HB(i2'#GH AxM(=>d0W~)f4F D4BRSD$Ljhw̉6nHw@PPuޞJme !w1V>a]M]oby”@׼bUu:*Y|FPU]oؔ[|G:ߪ!S6Jr)ؓ)SI.5o9oɼ+k`s&9dl ńuȎ}vhgTԘa)w0%'kFX+WC'mknug8C+\); ê??L83~kixn!yP_ݹ7V恿E1皰YqO!SJj?i*4=H!JOU Oi7p@zHů;UrjP&_C͖[CEtmݾe%QaVFtRkGi߻ $mvI!/fB5T9Io~*L:;I#Ļ}M9՝y.ޭ V>Љ}Fws! .vwk!1%-3i9oyWVdϔ:BǒP vI}*5CR zhQU FzhUR3RIñE)@.Vf (=lUjq?dJOz{E)Cx(4 )G6J C)! PJJ+ UǸQJJ)WOr(TbqDA1?2f0|.֤n#J~'X=I> _zZHM#JʳÌP<۳/!<QJ1薢aT!~^"?ӛ ݢʘ(u2^+@17Yh;'"TtWuZ^}Glv"i&Ri]b ԆFvBu+{W~vǨ#'9CKf1!?4;W1 ow.)ff 6=PQܭ>u0g_4"m$\Na$B$8/,]z0J; b!ᯧ_\P6X(ՠ{s&AnDpBXJ Iڶ|a(w#߆u3oa/d]ZSSNLb}ef $'-xk!˗3E8#yB E}v4At?EnN'o5AcPcjte:KlV,O焱QL"ly=p2iWdҝHRwg^z(cĝ|̓F0axO9y;wVOo63wmj-XcDW#%!]Yi9"x\b,NbtZ yf8hT'c&?|L236>~>waC>N +x.BȲ`HƤK.T^*.502"*;w N_:r,Ɗ}h#[a0ۂoʑsQD?9Ȑe~l?ݢWEv )g =hLX0gJQBT:Y Jm0'"/FJJqJ} JDWeȳ<Ւ4"+9Hd8U9…a4Q Esr9g-d7"dՎgczTR3w*כC609{}YuWSS꧋Kvy "Dw7zӫ|aU×'[T:q3IgM#|sPُWWppl0"daHT<~{tT2p31'ztSiwV 'e Y4-LzNGF J'Qbg_ RBfê>6vH%ʲ\A 6Ғa *`LU2M׶BT1燯BX3^2+Yg=*RV)QPVF B>YSB@sI/$Z`2?Ҫ6_ egR;z>l2RVqFI= 2JjQ=Qǚ pq-nq]W7.fn;枧Tnm^.FA+8%DV8- ޣQ->W{r2̶U!*T<LKd뻫|:%iw7&QhOy֦Z}NjG[xͨ?/-\cݷyp[ Y]}Je&t=qFmlpTOE496?vᡯ{ }TC٣%Ո$^$acĠknRhBz_JCPO L;-Ǝ?I?Y<[B!x1~_Wk ~~יy4m1 RH gLrOS3܂))ety駭]8SAFY"=?~2}dʿ{mX4Ϸxtr$ʿ>P V4~A݁eS)yٔW蜌 &CvfR&gkT}(Ǘuuf4P3z½GPGMwQT9BALҥX|EAx'/vО{pLKO'@ZaUf2 fPEm%4rzTkp tΊ{GK&$z|K߇UZuppv>!w&2޾}pGW{$cUI{_FQd{eۗBf]~p ߾6uw] *^<;{WØ*ۯQrpͽ32 9ۣQk$r(Rj}ݷ,ܨ?2jo)w]o SǓv==''؝EMGQb;EQgzVJFmTY Kky\o69?_eQ?;?/c &vw81dPS\( ̫1^* z(",X!:fa9B+FlU1O sgc&.V]󹅩5qgeʫ,uȋPKIٴ8a음@'qJޓt_5kd5 'g5еJ_hJZi^];ZQ%Ꟗ,?s/ᒍX*2CʣQA J#Kz*˲kKrQxnYV9Hl U^@7a%90Q8#tEenGAtHJ5Ļ`:D֙=F%7e,c/<*d, ycS"%&*e 4Y쵘`1mOXVI!^@Ě75 %LQ1BtEv d2`\!5~ ~$?^m.eظfzܒ*`$+Re9"-Z&^`.߫HeN_.I ȸI4vm"jPǚX >ODS59e=;i3nU<'UY%uJXٰa?mXW~ij|$(yMKF-r񤇱jNd.W4(nji%(lVRyWV: 3f|GO;cPOݕeHv&B6DE1;6;yy0dglFNE캦f45vœ85\D+vўuk3֚JG2 5LAjg"ȐC3\9;:c O']aL 6˚ /KhvKxIcaMx`̿Z-^JJ]s aZ<~(N<(۝;TV+h2k{9!Lt6eQ.S.&`ǟƢE܉'hVbN82"jm'CKңw'C?}LRݘ[l94 dʹ?~ XK28ytQc6n9rGbseNb:N1zµR6K@fZsp<81X 1F50gL< bU&L~d, ~(~c8+6mAnP2L <%ER]\L:"cN8veX1Lgsκ'&ob S  )cqYܩp="2|4: HĎs}F@.U?%i`ОCۉȞIH7eTWfn+d XW ?5KnbÎM#f6-S]$ E%B5U !6 vJGRϷFЏivZ|>_F)zTa]B@F&Kxr,6bYRlW7[ .S֪hS*^Dqނ3 t$Ʒe[67xN?D!N@5ٲ>RUc1 FusҧжxgoT#(a]s֝`=ZF1u* bVDtR[-A;occ٣s|-F!Mzr<<VulqQab;65B/ f pJ,0>i5FQt:\ۆ #`>٧@Y8*U"kC^]IVcqMLjW!ݙ ˧,D[+pHlɱqXWk\h~4\я_k^׼imB/'KK6jŘTcwh۲z-cyBsV00 ̟I5{T]h4JHHwE@2@@1݈ 57 dS!:~b?`SYloJkޒRˢ#PK |IJJv7A g>ߔ1 V{{IsTb*Pz:AjXT.r^ۦZ&W~vSL){Ek $ZZf .%jt^; ݕl&0җ1-Q:2h )-He+piƢVMoo[f}Mul{MΨ/uk{ WL-NJt&"JTp n"xVviP\s;c! ^ a/PP8T|ptB`$M8/ \ܮ-^, zJZZmi]VJ3cYCpkPJ]׫Vf<4n狠xUu"ہҦY:UҬ60T t:vf}a7^mNsfWjhb3^[M}c Ͱ K fJ3Rسk,aF[Qڴ0+ynʮJ[^)x1i)JВ,ux`|k1oMHy<˃;SZM՞r^4 p .VJI|ty.QVBhFc^I:P9zg`Eٻ6$W=MlRyEа`{q{^<-J<$:.ZMQf-,d~qefDz֠SrFר@qL(XU0ltHHS;Boؙ<,ajLI#*ЇAհXc"f/+)¹q`:}tx<ԾaYG:f{9W-R?n"}t*EsyJ\4LJ F1+cdō GX(1/!+AH16ry+?{BP2q:-h5F,Zf>HeBdxctݖr]?j}Go:jT02C+:k8+oٙ$]5.φGrȒxf>TAs<'#RRu蜼TC?Zo<TRnLqESYzxWZ/x_;KT] _enZ e3tF΂qSa+qA,YTX+(ơ[첱J>$ΧoP@3]aG[-a{-)mza6&Q'z:@* bP#M`Si=%DKJ,Ntp( ?U`੽ 䤙7AZ?<YKUawiBWO%QcВaiX⛄lJK-t`F 0.w$=#;X>>>*XȅeI;nHV"G`"v%{,w:N{5F5ZR־zǼ"EOߙy8"\p)ED gԬ*މѽ->aWf`dWQ睤YS )(D]Ku3T1TAPDѵG/RF ՙw¢;:5:+8A++%LW 3ww$jTL^10'3'ϔ.(0[{04Z:.Lj;˼(9DPYv墢+`-i0Ov"N6Uj΅Zh!FHgm;C︳wnù>:nBń^m^_eThR<CW_ygVud",~9h2eY1JEP?>Uiݑf;bZQa>m &j!0O9PE1tWM1$@MS ,j%| kk'l7_jGx?cU}O/N8bt@2cŹ`MNxkkngQa Uc!s^Mᰞ71Q2zn[*LTL^Z`m[!x7~QdrLl>QjmhaB*t+ƼʂZUCD -7v;/l |4 ] #~1X{$;Bٿ($-ʛd< At[0K@8Cek^{ʀ ;ax Wd=:ѲlP!2C5dB# "~:a}l5. Q .U7#:( iNB~hqETIk>)Xy)|kkKK=!oYߠ2JIPߜSG| Pλ_Ui:X GÈ$VC->G_ʧmBlI(.UM&duQHif?.gn.E0|clB$y`bWV8GK&عwgMՙ P=Z>U@h!eY{r:{쇚͠MbO̢P$(HgUa^X/cV!x-ۍV®NJ8FIfY1ioxbA\"ikm+ x;hj~'vȶRYOLf.^O|qR >E-^Z ējdm(c) ^Td9I6y Y`B9Ge NG^DoD夳}( @a}\TE"e;!.-Dܡ2:(<.4^# :EA#ctUW!oP2|@&n RH͆hnV8wjWn{ ӻTl<Z›v|b aP5-kVPsV!E ''keWig'ZgYy--բˉC[rJ Ey0Ea^-EU*cBAWJ4&bO UJ2jɷWX5s{B(Ϝk.OVRuR ̋TE"ശbmF=-,bBqs|%vic5$KΣN3s)r- i nZ˝ew7=j{Ee„i E IT UjGZzԙݑ'쭶v/c!\)ij :TۃU 9ػ(G@)R䙷NQ=ȂΉнS?ß6؎X;m.w"r{ 37 j%&V&%\iu~<8&gT^ĸ°p_\A,Ys} ~c^^ q*aiSE8vEF^ |!Ye4K%~#~{m9WEFe(ǩtq%ͻ˼f܆ >[#82?cbSa1H~fa]/q/lUEK^Eo2.+ XQv-/~^+vV t6⼯t`2~{Y*?tP?G|h:lT,!n_rU5 )ٌ}f^/ƃӳBc1.Y0?N,VtT|Ia|"Q>Bf~{YЀ_YYvΙu<Ơ2:by0mǠq xοSR@>xau̚ߍD4,< V2Ѵ_-7Rݏnb] jV0|i2 P`"(ǠIovYڞ ˍA‹~MW7 [pDPu@yJpriOq! 4߮^mpxb΋-d^ʤ`9 SջB\T.F&c^^\o&ې@-#B }~``kbg+.igDu@-,"Ÿh&iu^8(F;Q 1 XF|rQ (Mq?_ Ip6LqYxwޗӄ(oպb6Rא[. [V!Iы;1DK*>0R[PnFWKlIBoCTgϥP1-)(pCQ9;Bp[ )Nhﶡ"Ri,ȳf0Lf>Ŵ>}ؼpD?gW+> ~Kn+ܼq~DC[F'6:y'4Kއ_F yw6YB )wOF<> l~_xHx wYj*ip5kZl<Ʊhj|腲}e>N2 @3˙%`U;c)`kqD=H|v !ˇ]p]!AT*cͻ]S(h֥>_;m_kjίQv%([)>rVm EYb5:{j++Vk>[}čf멶*ʩ1ռB9AA݂.,6\hŞqf7mBkE2#ѱZ[=Z_c:F6ܥ4xJ68䕳h'Om0|𦰰/s9;qzny0rnAQMb3tQ +OHB6R6C%=n~SEy{TGKBbęK-[.<ř߭jBB$kH  X\r*zw¼8](?v֞YVT~TW!r?$,,1  `QJ 2L1FI*5ףϭ0t7kq11AM_t^4,Ch&K0:L;)q4]-9wғҕז!nMWq8փU i^H> D) zP$YD`Q$ +;[, ̸S "t";OMf&BC dV&dI#%h2_'7xŒ`b%ҮO& $i5:qh"UL7Di{kYIʠ)b*3̤y5c;e _ gzv>z@ u0hJ5^yg<"Zfx%۷&ظ4u(UwJ^x|!Sv蝲[M|&3OX2~: Yٳ4h] /!dsXг֘5YsI!:(1R',MEu uB}W΢A9CoU .ʻ:S)bd,NvU2Hpg,bh[!Srɸ9 O6ZtNMي*~#lOG)}ݰ79# Ŷh(Fo=2aXJ'?Y@T"UW̏}EpP>!ϛk6$\;yS"ŜꝹq z. Tׁ#.Jq9t\R*$Ԉ$Rc2S:_b.w+ s|}.I>!R_b\1aujLKI$PcE%p Nif%CiA8{b Pڷ]'[G(~gv#>b-u!GuysWu6e<8vmgX/jᓝa*}qϥO֤wM<lON"[d4C]_<;OTZpû3]I] ACt\YZ`y)d]K4w_rgrZVBWw}'#(i&^5 b)}w` 婮M*:C o9U %I5HcKL@XhIbpHEeS|p"U z.1ܱ;,UXjYIHe,R^W'+i5A~ϊ! ƃӲ^DԲx#*eut(`@=[:_)!1LXYm:I ZCˌ*YTaEL{} g~YJJE$zLj80 nU"щ1^kG<|jM!pGxa4@U]RVK!or-ۡÛ|1&IDZorM8s351H9%{Η!AT'$ ևkTt214:[eb(@i R,SP n(dH*D)OփfK#{g )"1<nͨ6 J3+AKL0JtF4Ȋ_ $Jpi2 0H240 Oweh5(5PJ F&Iw\KR):C^ݠJ$ F C1C j %ė+f#  jBeL2`?KX謤 :-=!֐hj$B $0" h5gȕUlWn\߷\!K,ӑ^:zS`r\~И,_믶Q7"%Ew ;.n~5B_ndebuWw!J)`ݷ"dMjCɼdԮ>16 ?Ž/'$CHpz8?+o3hmݣﴪ 5=vݭ'c3AJZRAh) &ZY{խ`2VzDŽ ؞|)(SpL>~>Y-ˍNBE\^:mlIdfA@OS6ss B!<#k, hɐ>sbJ;vb9%oVYFY12wIb!R!dǞIO69k?g!QyJ[x99o’&[ּqyXƱM57/P,X$Q,yD*NXgjMh⒬&3 ͬR(T |]림1^D:eirFrM3խ6p+܁<Ḕ:ouq8Tj KA 'L Α` R4IJ97'= ЮTxa'!2AH ={FADƈ@RnJa3RXJ $tF\Dbs\"SF%J"e={Pf+p*|-{ !&!BJ VqO͟GQi)GƧWV"9Mi{f=$5ll31'ByuW 1 WQD0ADb$wzW(?,7e`>Db";̤y$6yffmb޻"8.X:Pux=$S/yP"[!l^JX*v+h1if^Bqu0D\c㞛Ϛkp JY =]v}0kv2}JT춱,7Xtկ1r6zUaw6U$6KeVT!w貪~U!z~:Ʈݥ{e=󑜂Rr)ZTU鏭˕W+r!~Fܓ=:>U%y,S+zH7);넎QG N(MyJ68䕳hORιjMp~- t%Ow nmp+gk g/KH +#!pZI34Γfu ߤ8䕳hOҭǓ@a\ґnNuۀڗ)I𪄞ҭ y,S5CYnNuې T%+ {J68䕳h<2b:1Z3=? SDI)oz*z>r &+0˜vfwTv΀l!։-yx@;8TtΝel/{rk(|4.PF ^h^׫7vGݮVˋ ?Rp).NǛzLGOLm+?mly$ee6X} Psm}m9!EYW@Ʋtxǝ5nS[<ͼ/wt_E;5{1ry5qZ=Z4%+E1Zj'XŒ1b$mXFn^bgOǹs BA }:_'+=L|pΑ5VJ"-WB3+r$qB,32wƝţT@K5%-g5 TH&SӰ>R]fOmA?V+6SzMu8<<]Wf0Vf6&G5;&p~#^۵"|wNtlvmj :I+ITr|Aj >KK?~*'jzAZ| ʨ!Y'&R)&ލWЇjUS*?%M v!*>#~ 50岏hEJ]No-.arA‰ŀȋ%No>1|8/)"<bJ XFZh3R<G<řmKZrpADdaC-jPgKMaiBWKc 6,B*Yr6XqH0:}YɊSye{cwzP_uwH(z7ܿu-s7D:*+r& (z=hKƎRQHvV95J,@rR- (gF3FXHSXͪ ƽ8M= ǧQƏ*a IE QhNtp13rβBq̏1awxCXA<UF8Wg8k)[ay_Rhcwycٲeq,3lY2$(Jm_OJfg,2+ .-!rܗۂ(sJRx*3}NDcx_E&+_W#ވ+%Sfy :p`%apN`Ãp jʬjXU={T>x;) ܏n)͖r XA(lZ̤ Bqh瀻ܚϙ_9d2<' Q%ka&TF̬Xpx~0&(1QGf'u jxߑƘ1D_}H>ƶ"ԑ٣)9u9h;#ƺwgAj֭ E4I?zx׺UG9-AӦ5#{ wKo,: EtKxL#n đAaFܡvdŋ!ңȈ}mB{jJfvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005111716615146023256017711 0ustar rootrootFeb 20 08:05:49 crc systemd[1]: Starting Kubernetes Kubelet... Feb 20 08:05:49 crc restorecon[4679]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 08:05:49 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 08:05:50 crc restorecon[4679]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 20 08:05:51 crc kubenswrapper[4948]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 20 08:05:51 crc kubenswrapper[4948]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 20 08:05:51 crc kubenswrapper[4948]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 20 08:05:51 crc kubenswrapper[4948]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 20 08:05:51 crc kubenswrapper[4948]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 20 08:05:51 crc kubenswrapper[4948]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.449863 4948 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453454 4948 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453478 4948 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453485 4948 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453491 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453496 4948 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453501 4948 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453508 4948 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453514 4948 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453521 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453527 4948 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453533 4948 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453539 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453544 4948 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453549 4948 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453554 4948 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453560 4948 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453567 4948 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453573 4948 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453578 4948 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453583 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453588 4948 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453593 4948 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453598 4948 feature_gate.go:330] unrecognized feature gate: Example Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453603 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453609 4948 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453615 4948 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453620 4948 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453625 4948 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453631 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453636 4948 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453652 4948 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453658 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453663 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453669 4948 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453675 4948 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453680 4948 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453687 4948 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453692 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453697 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453702 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453707 4948 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453711 4948 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453716 4948 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453725 4948 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453730 4948 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453735 4948 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453740 4948 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453744 4948 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453749 4948 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453754 4948 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453759 4948 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453764 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453768 4948 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453773 4948 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453778 4948 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453782 4948 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453787 4948 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453792 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453796 4948 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453801 4948 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453806 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453810 4948 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453815 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453820 4948 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453825 4948 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453829 4948 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453835 4948 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453840 4948 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453844 4948 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453850 4948 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.453855 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.453957 4948 flags.go:64] FLAG: --address="0.0.0.0" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.453988 4948 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454002 4948 flags.go:64] FLAG: --anonymous-auth="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454010 4948 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454019 4948 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454024 4948 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454032 4948 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454039 4948 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454045 4948 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454051 4948 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454057 4948 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454063 4948 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454068 4948 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454074 4948 flags.go:64] FLAG: --cgroup-root="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454080 4948 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454085 4948 flags.go:64] FLAG: --client-ca-file="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454091 4948 flags.go:64] FLAG: --cloud-config="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454096 4948 flags.go:64] FLAG: --cloud-provider="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454102 4948 flags.go:64] FLAG: --cluster-dns="[]" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454135 4948 flags.go:64] FLAG: --cluster-domain="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454140 4948 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454146 4948 flags.go:64] FLAG: --config-dir="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454151 4948 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454158 4948 flags.go:64] FLAG: --container-log-max-files="5" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454166 4948 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454172 4948 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454178 4948 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454220 4948 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454227 4948 flags.go:64] FLAG: --contention-profiling="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454233 4948 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454239 4948 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454246 4948 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454252 4948 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454259 4948 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454265 4948 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454271 4948 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454277 4948 flags.go:64] FLAG: --enable-load-reader="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454282 4948 flags.go:64] FLAG: --enable-server="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454288 4948 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454294 4948 flags.go:64] FLAG: --event-burst="100" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454300 4948 flags.go:64] FLAG: --event-qps="50" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454306 4948 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454311 4948 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454317 4948 flags.go:64] FLAG: --eviction-hard="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454324 4948 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454330 4948 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454335 4948 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454341 4948 flags.go:64] FLAG: --eviction-soft="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454347 4948 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454354 4948 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454360 4948 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454366 4948 flags.go:64] FLAG: --experimental-mounter-path="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454371 4948 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454377 4948 flags.go:64] FLAG: --fail-swap-on="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454382 4948 flags.go:64] FLAG: --feature-gates="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454389 4948 flags.go:64] FLAG: --file-check-frequency="20s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454395 4948 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454401 4948 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454407 4948 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454413 4948 flags.go:64] FLAG: --healthz-port="10248" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454418 4948 flags.go:64] FLAG: --help="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454424 4948 flags.go:64] FLAG: --hostname-override="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454429 4948 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454435 4948 flags.go:64] FLAG: --http-check-frequency="20s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454440 4948 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454446 4948 flags.go:64] FLAG: --image-credential-provider-config="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454451 4948 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454458 4948 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454464 4948 flags.go:64] FLAG: --image-service-endpoint="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454469 4948 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454474 4948 flags.go:64] FLAG: --kube-api-burst="100" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454480 4948 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454486 4948 flags.go:64] FLAG: --kube-api-qps="50" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454491 4948 flags.go:64] FLAG: --kube-reserved="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454497 4948 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454502 4948 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454508 4948 flags.go:64] FLAG: --kubelet-cgroups="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454513 4948 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454519 4948 flags.go:64] FLAG: --lock-file="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454524 4948 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454530 4948 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454535 4948 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454544 4948 flags.go:64] FLAG: --log-json-split-stream="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454549 4948 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454555 4948 flags.go:64] FLAG: --log-text-split-stream="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454560 4948 flags.go:64] FLAG: --logging-format="text" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454566 4948 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454572 4948 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454577 4948 flags.go:64] FLAG: --manifest-url="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454583 4948 flags.go:64] FLAG: --manifest-url-header="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454590 4948 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454596 4948 flags.go:64] FLAG: --max-open-files="1000000" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454602 4948 flags.go:64] FLAG: --max-pods="110" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454608 4948 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454614 4948 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454620 4948 flags.go:64] FLAG: --memory-manager-policy="None" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454625 4948 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454631 4948 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454637 4948 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454642 4948 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454663 4948 flags.go:64] FLAG: --node-status-max-images="50" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454668 4948 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454674 4948 flags.go:64] FLAG: --oom-score-adj="-999" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454680 4948 flags.go:64] FLAG: --pod-cidr="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454686 4948 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454694 4948 flags.go:64] FLAG: --pod-manifest-path="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454699 4948 flags.go:64] FLAG: --pod-max-pids="-1" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454706 4948 flags.go:64] FLAG: --pods-per-core="0" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454712 4948 flags.go:64] FLAG: --port="10250" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454717 4948 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454723 4948 flags.go:64] FLAG: --provider-id="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454728 4948 flags.go:64] FLAG: --qos-reserved="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454734 4948 flags.go:64] FLAG: --read-only-port="10255" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454740 4948 flags.go:64] FLAG: --register-node="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454745 4948 flags.go:64] FLAG: --register-schedulable="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454750 4948 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454760 4948 flags.go:64] FLAG: --registry-burst="10" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454765 4948 flags.go:64] FLAG: --registry-qps="5" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454771 4948 flags.go:64] FLAG: --reserved-cpus="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454776 4948 flags.go:64] FLAG: --reserved-memory="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454783 4948 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454789 4948 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454795 4948 flags.go:64] FLAG: --rotate-certificates="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454800 4948 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454806 4948 flags.go:64] FLAG: --runonce="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454811 4948 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454817 4948 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454822 4948 flags.go:64] FLAG: --seccomp-default="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454828 4948 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454834 4948 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454839 4948 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454845 4948 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454850 4948 flags.go:64] FLAG: --storage-driver-password="root" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454858 4948 flags.go:64] FLAG: --storage-driver-secure="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454865 4948 flags.go:64] FLAG: --storage-driver-table="stats" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454870 4948 flags.go:64] FLAG: --storage-driver-user="root" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454875 4948 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454882 4948 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454887 4948 flags.go:64] FLAG: --system-cgroups="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454894 4948 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454903 4948 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454908 4948 flags.go:64] FLAG: --tls-cert-file="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454913 4948 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454920 4948 flags.go:64] FLAG: --tls-min-version="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454926 4948 flags.go:64] FLAG: --tls-private-key-file="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454932 4948 flags.go:64] FLAG: --topology-manager-policy="none" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454937 4948 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454943 4948 flags.go:64] FLAG: --topology-manager-scope="container" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454948 4948 flags.go:64] FLAG: --v="2" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454956 4948 flags.go:64] FLAG: --version="false" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454963 4948 flags.go:64] FLAG: --vmodule="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454986 4948 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.454993 4948 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455157 4948 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455164 4948 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455170 4948 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455176 4948 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455182 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455187 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455195 4948 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455200 4948 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455205 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455210 4948 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455215 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455219 4948 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455227 4948 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455231 4948 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455236 4948 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455241 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455246 4948 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455251 4948 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455256 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455260 4948 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455265 4948 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455270 4948 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455276 4948 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455281 4948 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455286 4948 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455290 4948 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455295 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455300 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455305 4948 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455310 4948 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455315 4948 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455319 4948 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455324 4948 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455329 4948 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455334 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455340 4948 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455347 4948 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455353 4948 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455360 4948 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455366 4948 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455371 4948 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455376 4948 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455381 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455386 4948 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455393 4948 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455397 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455403 4948 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455409 4948 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455415 4948 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455420 4948 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455426 4948 feature_gate.go:330] unrecognized feature gate: Example Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455432 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455437 4948 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455442 4948 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455446 4948 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455451 4948 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455456 4948 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455460 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455466 4948 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455471 4948 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455477 4948 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455483 4948 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455490 4948 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455496 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455502 4948 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455507 4948 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455513 4948 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455518 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455523 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455528 4948 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.455534 4948 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.456259 4948 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.468425 4948 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.468883 4948 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469037 4948 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469053 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469063 4948 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469072 4948 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469082 4948 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469090 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469098 4948 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469106 4948 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469113 4948 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469124 4948 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469138 4948 feature_gate.go:330] unrecognized feature gate: Example Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469148 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469157 4948 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469166 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469175 4948 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469183 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469191 4948 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469201 4948 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469209 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469216 4948 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469224 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469232 4948 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469240 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469248 4948 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469257 4948 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469267 4948 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469277 4948 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469287 4948 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469296 4948 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469307 4948 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469318 4948 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469327 4948 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469335 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469344 4948 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469354 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469362 4948 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469371 4948 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469379 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469386 4948 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469394 4948 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469402 4948 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469411 4948 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469419 4948 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469427 4948 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469435 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469443 4948 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469451 4948 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469459 4948 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469466 4948 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469474 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469482 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469490 4948 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469498 4948 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469506 4948 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469515 4948 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469523 4948 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469531 4948 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469539 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469547 4948 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469554 4948 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469562 4948 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469570 4948 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469577 4948 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469585 4948 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469592 4948 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469600 4948 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469608 4948 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469616 4948 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469626 4948 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469636 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469646 4948 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.469660 4948 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469879 4948 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469893 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469902 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469911 4948 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469920 4948 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469929 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469936 4948 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469944 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469952 4948 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469959 4948 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469967 4948 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.469997 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470005 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470013 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470021 4948 feature_gate.go:330] unrecognized feature gate: Example Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470029 4948 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470041 4948 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470048 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470056 4948 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470064 4948 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470072 4948 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470079 4948 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470087 4948 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470095 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470102 4948 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470110 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470118 4948 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470125 4948 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470133 4948 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470141 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470151 4948 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470162 4948 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470170 4948 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470178 4948 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470187 4948 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470195 4948 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470205 4948 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470215 4948 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470225 4948 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470233 4948 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470243 4948 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470252 4948 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470260 4948 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470268 4948 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470276 4948 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470284 4948 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470292 4948 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470300 4948 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470308 4948 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470316 4948 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470323 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470331 4948 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470339 4948 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470346 4948 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470354 4948 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470362 4948 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470369 4948 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470380 4948 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470388 4948 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470398 4948 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470406 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470414 4948 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470422 4948 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470431 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470439 4948 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470447 4948 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470455 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470463 4948 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470470 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470477 4948 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.470486 4948 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.470498 4948 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.471560 4948 server.go:940] "Client rotation is on, will bootstrap in background" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.483304 4948 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.483507 4948 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.485464 4948 server.go:997] "Starting client certificate rotation" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.485531 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.485821 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-15 14:17:16.878034339 +0000 UTC Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.486006 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.511641 4948 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.516409 4948 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.518438 4948 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.542550 4948 log.go:25] "Validated CRI v1 runtime API" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.583288 4948 log.go:25] "Validated CRI v1 image API" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.585935 4948 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.591651 4948 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-20-08-01-26-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.591698 4948 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.619030 4948 manager.go:217] Machine: {Timestamp:2026-02-20 08:05:51.615328761 +0000 UTC m=+0.589823641 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654132736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:51d89745-b91f-4e85-9724-1ef53e3862a9 BootID:8f36346c-e92f-4a00-a0af-e0652f71277d Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730829824 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827068416 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:85:61:b1 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:85:61:b1 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:e8:40:45 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:13:c9:98 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:bf:26:78 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:f4:2b:94 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:7e:8d:3c:94:36:7d Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:8e:8f:b6:78:f7:89 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654132736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.619446 4948 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.619680 4948 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.623355 4948 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.623744 4948 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.623856 4948 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.625332 4948 topology_manager.go:138] "Creating topology manager with none policy" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.625370 4948 container_manager_linux.go:303] "Creating device plugin manager" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.626036 4948 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.626113 4948 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.626705 4948 state_mem.go:36] "Initialized new in-memory state store" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.626963 4948 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.631006 4948 kubelet.go:418] "Attempting to sync node with API server" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.631089 4948 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.631186 4948 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.631212 4948 kubelet.go:324] "Adding apiserver pod source" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.631233 4948 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.636245 4948 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.637543 4948 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.637890 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.637941 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.638084 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.638084 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.640460 4948 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642424 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642467 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642483 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642496 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642518 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642531 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642545 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642568 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642585 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642600 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642622 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.642635 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.643889 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.644615 4948 server.go:1280] "Started kubelet" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.646238 4948 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.646284 4948 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.646519 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.647071 4948 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 20 08:05:51 crc systemd[1]: Started Kubernetes Kubelet. Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.648924 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.649336 4948 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.649375 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 06:37:54.603984492 +0000 UTC Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.649565 4948 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.649586 4948 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.649639 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.649682 4948 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.650826 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="200ms" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.651391 4948 server.go:460] "Adding debug handlers to kubelet server" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.651464 4948 factory.go:55] Registering systemd factory Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.651501 4948 factory.go:221] Registration of the systemd container factory successfully Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.652075 4948 factory.go:153] Registering CRI-O factory Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.652103 4948 factory.go:221] Registration of the crio container factory successfully Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.652234 4948 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.652276 4948 factory.go:103] Registering Raw factory Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.652303 4948 manager.go:1196] Started watching for new ooms in manager Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.652344 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.652488 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.653403 4948 manager.go:319] Starting recovery of all containers Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.652828 4948 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.113:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1895e5d9b4d34f97 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-20 08:05:51.644569495 +0000 UTC m=+0.619064345,LastTimestamp:2026-02-20 08:05:51.644569495 +0000 UTC m=+0.619064345,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676451 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676519 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676544 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676563 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676583 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676603 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676622 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676641 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676663 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676683 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676751 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676772 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676791 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676834 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676854 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676874 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676903 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676923 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676940 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.676960 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677014 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677041 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677064 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677132 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677153 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677172 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677195 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677217 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677235 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677257 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677275 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677296 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677316 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677335 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677355 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677374 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677392 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677447 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677465 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677483 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677501 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677518 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677537 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677554 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677577 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677596 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677616 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677637 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677656 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677678 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677699 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677718 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677743 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.677766 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.679945 4948 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680053 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680087 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680109 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680132 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680153 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680177 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680199 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680218 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680240 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680259 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680278 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680297 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680316 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680335 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680354 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680372 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680392 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680412 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680432 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680471 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680491 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680510 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680529 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680545 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680569 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680589 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680609 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680627 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680646 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680666 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680685 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680703 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680722 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680740 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680757 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680775 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680793 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680811 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680829 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680849 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680872 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680894 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680914 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680933 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.680954 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681011 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681037 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681058 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681078 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681134 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681163 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681184 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681205 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681229 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681250 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681272 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681294 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681315 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681335 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681357 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681377 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681395 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681414 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681432 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681451 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681467 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681485 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681504 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681522 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681540 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681563 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681580 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681598 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681617 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681637 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681656 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681675 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681697 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681714 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681731 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681749 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681768 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681787 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681804 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681822 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681840 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681865 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681882 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681901 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681922 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681940 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.681958 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682010 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682034 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682052 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682070 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682088 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682105 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682124 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682142 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682159 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682177 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682195 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682212 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682229 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682247 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682263 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682280 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682297 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682315 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682335 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682353 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682371 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682389 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682405 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682425 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682444 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682461 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682484 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682502 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682535 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682555 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682574 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682591 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682610 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682629 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682648 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682668 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682696 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682714 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682732 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682749 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682766 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682785 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682803 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682821 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682841 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682859 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682877 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682897 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682915 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682934 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682952 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.682996 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683055 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683078 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683097 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683113 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683134 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683152 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683177 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683194 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683215 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683233 4948 reconstruct.go:97] "Volume reconstruction finished" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.683246 4948 reconciler.go:26] "Reconciler: start to sync state" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.692575 4948 manager.go:324] Recovery completed Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.711788 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.713964 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.714172 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.714269 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.715469 4948 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.715598 4948 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.715683 4948 state_mem.go:36] "Initialized new in-memory state store" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.717181 4948 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.721053 4948 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.721146 4948 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.721216 4948 kubelet.go:2335] "Starting kubelet main sync loop" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.721308 4948 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 20 08:05:51 crc kubenswrapper[4948]: W0220 08:05:51.722199 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.722302 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.739348 4948 policy_none.go:49] "None policy: Start" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.741526 4948 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.741664 4948 state_mem.go:35] "Initializing new in-memory state store" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.749762 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.811583 4948 manager.go:334] "Starting Device Plugin manager" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.811937 4948 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.811968 4948 server.go:79] "Starting device plugin registration server" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.812582 4948 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.812612 4948 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.813587 4948 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.813790 4948 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.813807 4948 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.821802 4948 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.821950 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.824489 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.824572 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.824597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.824880 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.825249 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.825322 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.830784 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.830788 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.830827 4948 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.830857 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.830870 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.830890 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.830893 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.831966 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.832265 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.832361 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.834269 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.834314 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.834268 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.834364 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.834383 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.834334 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.834639 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.834879 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.835007 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.835904 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.835945 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.836005 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.836155 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.836279 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.836318 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.836336 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.836756 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.836813 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.837581 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.837632 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.837653 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.837850 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.837885 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.838645 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.838685 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.838704 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.839395 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.839451 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.839473 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.851707 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="400ms" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.885481 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.886009 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.886448 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.886538 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.886729 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.887095 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.887325 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.887690 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.887907 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.888270 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.888487 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.888674 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.888851 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.888961 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.889043 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.913359 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.915026 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.915095 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.915117 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.915164 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 08:05:51 crc kubenswrapper[4948]: E0220 08:05:51.916203 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.113:6443: connect: connection refused" node="crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.990693 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.990755 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.990804 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.990840 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.990912 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.990944 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991007 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991006 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991106 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991100 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991104 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991038 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991216 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991270 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991016 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991206 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991243 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991259 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991370 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991408 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991439 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991469 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991485 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991518 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991526 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991501 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991590 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991553 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991655 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:51 crc kubenswrapper[4948]: I0220 08:05:51.991555 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.116635 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.119124 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.119190 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.119211 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.119255 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 08:05:52 crc kubenswrapper[4948]: E0220 08:05:52.119769 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.113:6443: connect: connection refused" node="crc" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.179856 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.195088 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.211650 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.233038 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:52 crc kubenswrapper[4948]: W0220 08:05:52.238014 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-fb97570385558360c8e37ed85db7ebad3ca89392dbdd9304506115ac70cfcc6a WatchSource:0}: Error finding container fb97570385558360c8e37ed85db7ebad3ca89392dbdd9304506115ac70cfcc6a: Status 404 returned error can't find the container with id fb97570385558360c8e37ed85db7ebad3ca89392dbdd9304506115ac70cfcc6a Feb 20 08:05:52 crc kubenswrapper[4948]: W0220 08:05:52.240089 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-7dd21173ef20d7b33a38877aa593fa065ed43716bf18e03a37a95cf27c3b5b2d WatchSource:0}: Error finding container 7dd21173ef20d7b33a38877aa593fa065ed43716bf18e03a37a95cf27c3b5b2d: Status 404 returned error can't find the container with id 7dd21173ef20d7b33a38877aa593fa065ed43716bf18e03a37a95cf27c3b5b2d Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.242620 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 08:05:52 crc kubenswrapper[4948]: E0220 08:05:52.252675 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="800ms" Feb 20 08:05:52 crc kubenswrapper[4948]: W0220 08:05:52.254516 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-e5cb1d98ca7c049a62301d2cdb2cadea4ba1ba0b36b390cd5329fb1b233ddfb4 WatchSource:0}: Error finding container e5cb1d98ca7c049a62301d2cdb2cadea4ba1ba0b36b390cd5329fb1b233ddfb4: Status 404 returned error can't find the container with id e5cb1d98ca7c049a62301d2cdb2cadea4ba1ba0b36b390cd5329fb1b233ddfb4 Feb 20 08:05:52 crc kubenswrapper[4948]: W0220 08:05:52.261089 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-efec26f7c799b90d098d9b607041426a10d2dfaa0c9145be7330d16b271ea176 WatchSource:0}: Error finding container efec26f7c799b90d098d9b607041426a10d2dfaa0c9145be7330d16b271ea176: Status 404 returned error can't find the container with id efec26f7c799b90d098d9b607041426a10d2dfaa0c9145be7330d16b271ea176 Feb 20 08:05:52 crc kubenswrapper[4948]: W0220 08:05:52.263054 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-090a66e9126788a7d8e207e7d0a743e706639096359b71eab9813e032dffe803 WatchSource:0}: Error finding container 090a66e9126788a7d8e207e7d0a743e706639096359b71eab9813e032dffe803: Status 404 returned error can't find the container with id 090a66e9126788a7d8e207e7d0a743e706639096359b71eab9813e032dffe803 Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.520585 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.521875 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.521938 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.521957 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.522032 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 08:05:52 crc kubenswrapper[4948]: E0220 08:05:52.522714 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.113:6443: connect: connection refused" node="crc" Feb 20 08:05:52 crc kubenswrapper[4948]: W0220 08:05:52.592119 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:52 crc kubenswrapper[4948]: E0220 08:05:52.592270 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:52 crc kubenswrapper[4948]: W0220 08:05:52.612332 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:52 crc kubenswrapper[4948]: E0220 08:05:52.612443 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.648273 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.650481 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 14:19:00.424502236 +0000 UTC Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.728635 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"090a66e9126788a7d8e207e7d0a743e706639096359b71eab9813e032dffe803"} Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.730468 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"efec26f7c799b90d098d9b607041426a10d2dfaa0c9145be7330d16b271ea176"} Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.731611 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e5cb1d98ca7c049a62301d2cdb2cadea4ba1ba0b36b390cd5329fb1b233ddfb4"} Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.733156 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7dd21173ef20d7b33a38877aa593fa065ed43716bf18e03a37a95cf27c3b5b2d"} Feb 20 08:05:52 crc kubenswrapper[4948]: I0220 08:05:52.734501 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fb97570385558360c8e37ed85db7ebad3ca89392dbdd9304506115ac70cfcc6a"} Feb 20 08:05:52 crc kubenswrapper[4948]: W0220 08:05:52.941380 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:52 crc kubenswrapper[4948]: E0220 08:05:52.941502 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:53 crc kubenswrapper[4948]: E0220 08:05:53.054386 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="1.6s" Feb 20 08:05:53 crc kubenswrapper[4948]: W0220 08:05:53.119819 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:53 crc kubenswrapper[4948]: E0220 08:05:53.119963 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.323402 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.325018 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.325054 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.325064 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.325087 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 08:05:53 crc kubenswrapper[4948]: E0220 08:05:53.325615 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.113:6443: connect: connection refused" node="crc" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.593156 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 20 08:05:53 crc kubenswrapper[4948]: E0220 08:05:53.595038 4948 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.647628 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.650723 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 20:09:05.356180009 +0000 UTC Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.743323 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378"} Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.743395 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737"} Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.743418 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4"} Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.745533 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b" exitCode=0 Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.745699 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.745954 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b"} Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.747208 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.747258 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.747277 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.748547 4948 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe" exitCode=0 Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.748614 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe"} Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.748739 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.750232 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.750412 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.750459 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.750474 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.751408 4948 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221" exitCode=0 Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.751480 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221"} Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.751583 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.752739 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.752784 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.752802 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.753204 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.753279 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.753307 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.754758 4948 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054" exitCode=0 Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.754828 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054"} Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.755049 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.756289 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.756315 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:53 crc kubenswrapper[4948]: I0220 08:05:53.756344 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:54 crc kubenswrapper[4948]: W0220 08:05:54.506076 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:54 crc kubenswrapper[4948]: E0220 08:05:54.506611 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.647486 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.650904 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 22:14:02.857179738 +0000 UTC Feb 20 08:05:54 crc kubenswrapper[4948]: E0220 08:05:54.655493 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="3.2s" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.767831 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.767987 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.769341 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.769372 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.769381 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.774015 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.774707 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.774764 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.774784 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.776650 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.776696 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.776726 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.779671 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.779739 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.780892 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.780923 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.780934 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.783648 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.783675 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.783687 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.783697 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.793230 4948 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc" exitCode=0 Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.793304 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc"} Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.793402 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.794737 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.794771 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.794789 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:54 crc kubenswrapper[4948]: E0220 08:05:54.874745 4948 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.113:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1895e5d9b4d34f97 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-20 08:05:51.644569495 +0000 UTC m=+0.619064345,LastTimestamp:2026-02-20 08:05:51.644569495 +0000 UTC m=+0.619064345,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.926041 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.927284 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.927324 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.927334 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.927362 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 08:05:54 crc kubenswrapper[4948]: E0220 08:05:54.927943 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.113:6443: connect: connection refused" node="crc" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.970038 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:54 crc kubenswrapper[4948]: I0220 08:05:54.979744 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:05:55 crc kubenswrapper[4948]: W0220 08:05:55.015425 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:55 crc kubenswrapper[4948]: E0220 08:05:55.015515 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:55 crc kubenswrapper[4948]: W0220 08:05:55.058399 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.113:6443: connect: connection refused Feb 20 08:05:55 crc kubenswrapper[4948]: E0220 08:05:55.058512 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.113:6443: connect: connection refused" logger="UnhandledError" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.329238 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.651806 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 14:49:40.557334232 +0000 UTC Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.799625 4948 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb" exitCode=0 Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.799783 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb"} Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.799808 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.801331 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.801380 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.801402 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.804256 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678"} Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.804340 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.804382 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.804393 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.806476 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811385 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811423 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811432 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811431 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811466 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811468 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811507 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811479 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811538 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811569 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811588 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:55 crc kubenswrapper[4948]: I0220 08:05:55.811525 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.653075 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 03:35:14.409499108 +0000 UTC Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.812854 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.812910 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.812927 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.813041 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.813470 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.813722 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e"} Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.813798 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7"} Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.813820 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154"} Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.814208 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.814256 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.814272 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.815153 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.815213 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.815232 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.815479 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.815554 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:56 crc kubenswrapper[4948]: I0220 08:05:56.815577 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.295206 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.340632 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.654138 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 12:28:10.201801422 +0000 UTC Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.701359 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.822648 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5"} Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.822740 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc"} Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.822681 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.822682 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.824413 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.824484 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.824491 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.824598 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.824534 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:57 crc kubenswrapper[4948]: I0220 08:05:57.824655 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.128557 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.130131 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.130199 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.130222 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.130266 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.307940 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.654781 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 17:20:54.709280079 +0000 UTC Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.825923 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.825923 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.827204 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.827791 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.828040 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.828104 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.828054 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:58 crc kubenswrapper[4948]: I0220 08:05:58.828206 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:05:59 crc kubenswrapper[4948]: I0220 08:05:59.655247 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 06:23:03.129009847 +0000 UTC Feb 20 08:05:59 crc kubenswrapper[4948]: I0220 08:05:59.829307 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:05:59 crc kubenswrapper[4948]: I0220 08:05:59.830641 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:05:59 crc kubenswrapper[4948]: I0220 08:05:59.830711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:05:59 crc kubenswrapper[4948]: I0220 08:05:59.830737 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.245285 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.245525 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.245595 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.247433 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.247635 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.247757 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.498323 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.498642 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.500505 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.500571 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.500592 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:00 crc kubenswrapper[4948]: I0220 08:06:00.656481 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 05:08:56.78082992 +0000 UTC Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.254390 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.254647 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.256569 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.256627 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.256647 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.657010 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 16:35:00.138202598 +0000 UTC Feb 20 08:06:01 crc kubenswrapper[4948]: E0220 08:06:01.831406 4948 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.985063 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.985346 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.987071 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.987136 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:01 crc kubenswrapper[4948]: I0220 08:06:01.987159 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:02 crc kubenswrapper[4948]: I0220 08:06:02.657358 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 00:07:42.351053568 +0000 UTC Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.658469 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 00:48:14.890162779 +0000 UTC Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.838658 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.839013 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.841290 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.841360 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.841377 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.847694 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.866898 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.868420 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.868467 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:03 crc kubenswrapper[4948]: I0220 08:06:03.868486 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:04 crc kubenswrapper[4948]: I0220 08:06:04.658700 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 08:29:25.256040411 +0000 UTC Feb 20 08:06:05 crc kubenswrapper[4948]: W0220 08:06:05.321899 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 20 08:06:05 crc kubenswrapper[4948]: I0220 08:06:05.322109 4948 trace.go:236] Trace[904507372]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Feb-2026 08:05:55.320) (total time: 10001ms): Feb 20 08:06:05 crc kubenswrapper[4948]: Trace[904507372]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (08:06:05.321) Feb 20 08:06:05 crc kubenswrapper[4948]: Trace[904507372]: [10.001802632s] [10.001802632s] END Feb 20 08:06:05 crc kubenswrapper[4948]: E0220 08:06:05.322150 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 20 08:06:05 crc kubenswrapper[4948]: I0220 08:06:05.648844 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 20 08:06:05 crc kubenswrapper[4948]: I0220 08:06:05.659051 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 12:45:04.669657577 +0000 UTC Feb 20 08:06:06 crc kubenswrapper[4948]: I0220 08:06:06.659652 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 21:10:33.544935053 +0000 UTC Feb 20 08:06:06 crc kubenswrapper[4948]: I0220 08:06:06.839386 4948 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 20 08:06:06 crc kubenswrapper[4948]: I0220 08:06:06.839505 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 20 08:06:06 crc kubenswrapper[4948]: I0220 08:06:06.976350 4948 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 20 08:06:06 crc kubenswrapper[4948]: I0220 08:06:06.976472 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 20 08:06:06 crc kubenswrapper[4948]: I0220 08:06:06.984816 4948 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Feb 20 08:06:06 crc kubenswrapper[4948]: I0220 08:06:06.984873 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 20 08:06:07 crc kubenswrapper[4948]: I0220 08:06:07.350501 4948 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]log ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]etcd ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/openshift.io-api-request-count-filter ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/openshift.io-startkubeinformers ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/generic-apiserver-start-informers ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/priority-and-fairness-config-consumer ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/priority-and-fairness-filter ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-apiextensions-informers ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-apiextensions-controllers ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/crd-informer-synced ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-system-namespaces-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-cluster-authentication-info-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-legacy-token-tracking-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-service-ip-repair-controllers ok Feb 20 08:06:07 crc kubenswrapper[4948]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Feb 20 08:06:07 crc kubenswrapper[4948]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/priority-and-fairness-config-producer ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/bootstrap-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/start-kube-aggregator-informers ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/apiservice-status-local-available-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/apiservice-status-remote-available-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/apiservice-registration-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/apiservice-wait-for-first-sync ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/apiservice-discovery-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/kube-apiserver-autoregistration ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]autoregister-completion ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/apiservice-openapi-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: [+]poststarthook/apiservice-openapiv3-controller ok Feb 20 08:06:07 crc kubenswrapper[4948]: livez check failed Feb 20 08:06:07 crc kubenswrapper[4948]: I0220 08:06:07.350601 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:06:07 crc kubenswrapper[4948]: I0220 08:06:07.660085 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 02:32:07.650627102 +0000 UTC Feb 20 08:06:08 crc kubenswrapper[4948]: I0220 08:06:08.660480 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 13:51:37.626336664 +0000 UTC Feb 20 08:06:09 crc kubenswrapper[4948]: I0220 08:06:09.661592 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 22:08:24.341362603 +0000 UTC Feb 20 08:06:10 crc kubenswrapper[4948]: I0220 08:06:10.662235 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 09:32:00.608275077 +0000 UTC Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.087836 4948 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.662456 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 22:21:46.593300688 +0000 UTC Feb 20 08:06:11 crc kubenswrapper[4948]: E0220 08:06:11.831708 4948 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 20 08:06:11 crc kubenswrapper[4948]: E0220 08:06:11.980201 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.985499 4948 trace.go:236] Trace[1559600258]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Feb-2026 08:05:58.510) (total time: 13475ms): Feb 20 08:06:11 crc kubenswrapper[4948]: Trace[1559600258]: ---"Objects listed" error: 13475ms (08:06:11.985) Feb 20 08:06:11 crc kubenswrapper[4948]: Trace[1559600258]: [13.475421278s] [13.475421278s] END Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.985544 4948 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 20 08:06:11 crc kubenswrapper[4948]: E0220 08:06:11.988019 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.988380 4948 trace.go:236] Trace[423861376]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Feb-2026 08:05:59.591) (total time: 12397ms): Feb 20 08:06:11 crc kubenswrapper[4948]: Trace[423861376]: ---"Objects listed" error: 12397ms (08:06:11.988) Feb 20 08:06:11 crc kubenswrapper[4948]: Trace[423861376]: [12.397093228s] [12.397093228s] END Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.988404 4948 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.988401 4948 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.990238 4948 trace.go:236] Trace[877424968]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Feb-2026 08:05:59.335) (total time: 12654ms): Feb 20 08:06:11 crc kubenswrapper[4948]: Trace[877424968]: ---"Objects listed" error: 12654ms (08:06:11.990) Feb 20 08:06:11 crc kubenswrapper[4948]: Trace[877424968]: [12.654435359s] [12.654435359s] END Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.990261 4948 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 20 08:06:11 crc kubenswrapper[4948]: I0220 08:06:11.991914 4948 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.021130 4948 csr.go:261] certificate signing request csr-s6842 is approved, waiting to be issued Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.032521 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.035907 4948 csr.go:257] certificate signing request csr-s6842 is issued Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.041562 4948 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42940->192.168.126.11:17697: read: connection reset by peer" start-of-body= Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.041652 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42940->192.168.126.11:17697: read: connection reset by peer" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.041565 4948 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42930->192.168.126.11:17697: read: connection reset by peer" start-of-body= Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.041797 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42930->192.168.126.11:17697: read: connection reset by peer" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.072810 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.349345 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.350553 4948 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.350678 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.356507 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.646914 4948 apiserver.go:52] "Watching apiserver" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.650419 4948 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.650723 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.651066 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.651190 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.651366 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.651500 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.651513 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.651557 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.651607 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.651796 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.651894 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.653200 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.653222 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.654002 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.654134 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.654641 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.654654 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.655039 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.656245 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.656474 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.663432 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 11:42:50.465544472 +0000 UTC Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.682227 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.694163 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.711057 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.726118 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.739686 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.750551 4948 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.751906 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.773743 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795008 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795056 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795074 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795094 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795112 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795130 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795149 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795167 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795189 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795212 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795234 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795254 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795271 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795287 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795302 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795319 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795335 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795350 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795405 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795424 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795443 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795458 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795474 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795492 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795509 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795530 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795552 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795578 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795601 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795623 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795641 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.795921 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796140 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796413 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796458 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796465 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796498 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796539 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796690 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796738 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796833 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796869 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.796968 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797037 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797068 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797080 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797109 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797154 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797215 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797340 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797423 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797471 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.797582 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798092 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798297 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798334 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798362 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798385 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798411 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798434 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798458 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798483 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798508 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798529 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798879 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799083 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.798530 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799416 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799451 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799480 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799505 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799532 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799560 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799595 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799595 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799643 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799702 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799735 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799759 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799770 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799817 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799834 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799858 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799881 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799902 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799927 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799949 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.799993 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800019 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800040 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800066 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800089 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800114 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800136 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800157 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800175 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800196 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800215 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800236 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800257 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800279 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800301 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800317 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800334 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800332 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800350 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800445 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800458 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800445 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800681 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800693 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800785 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800813 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800820 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800866 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800889 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800933 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.800967 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801032 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801066 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801098 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801131 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801140 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801163 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801200 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801256 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801281 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801291 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801335 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801366 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801398 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801431 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801464 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801497 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801534 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801567 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801618 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801656 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801690 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801723 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801764 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801808 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801850 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801896 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801930 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801961 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802020 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802055 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802086 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802117 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802151 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802181 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802215 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802250 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802283 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802314 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802349 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802384 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802416 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802450 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802485 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802519 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802553 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802590 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802622 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802657 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802692 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802725 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802761 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802793 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802825 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802857 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802891 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802923 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802961 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803137 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803194 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803602 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803641 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803679 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803713 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803750 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803785 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803822 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803855 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803889 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803924 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803957 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804020 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804420 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804470 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804506 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804540 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804675 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804714 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804750 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804702 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804785 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804819 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804853 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804887 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804922 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804953 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805010 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805048 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805082 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805117 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805190 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805227 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805263 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805299 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805335 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805368 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805401 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805434 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805471 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805508 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805542 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805573 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805608 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805642 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805675 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805710 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.805744 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807217 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807408 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807451 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807490 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807574 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807715 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807752 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807788 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807827 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807863 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807896 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808136 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808194 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808238 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808274 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808314 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808352 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808394 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808432 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808467 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808512 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808591 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801491 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808646 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801522 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801539 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.801806 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802154 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802411 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802572 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802744 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.802936 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808781 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803318 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803665 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803852 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803944 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.803963 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804083 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804168 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804343 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.804732 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807192 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.807542 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808103 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808182 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808374 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808712 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808943 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.808770 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809095 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809329 4948 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809348 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809363 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809377 4948 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809392 4948 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809404 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809419 4948 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809436 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809451 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809466 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809479 4948 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809491 4948 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809503 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809517 4948 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809531 4948 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809545 4948 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809558 4948 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809572 4948 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809586 4948 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809599 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809613 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809630 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809643 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809657 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809668 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809681 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809694 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809707 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809720 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809732 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809743 4948 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809757 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809769 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809781 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809792 4948 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809804 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809816 4948 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809828 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809840 4948 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809853 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809865 4948 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809877 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809891 4948 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809903 4948 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809916 4948 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809928 4948 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809940 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809952 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809964 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809996 4948 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809437 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809446 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809483 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809509 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809577 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.809812 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810000 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810097 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.810220 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810223 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.810288 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:13.310266656 +0000 UTC m=+22.284761606 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810380 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810347 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810455 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810843 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810965 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.810958 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.811287 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.811271 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.811598 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.811876 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.811906 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.812081 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.812319 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.812545 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.812737 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.812780 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.812796 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.812152 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.812858 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.813102 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.813181 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.813204 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.813662 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.813582 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.813829 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.814031 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.814317 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.814378 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.814499 4948 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.814706 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.814796 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.814822 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.814989 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.815200 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.815155 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.817361 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.817476 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:13.317443217 +0000 UTC m=+22.291938077 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.817927 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.818253 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.818461 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.821466 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:06:13.321442968 +0000 UTC m=+22.295937778 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.821455 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.830291 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.830670 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.830671 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.831777 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.833155 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.833201 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.833221 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.833668 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.833863 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.834033 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:13.334005295 +0000 UTC m=+22.308500325 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.834266 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.834354 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.834460 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.834588 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:13.334569619 +0000 UTC m=+22.309064439 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.834299 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.834286 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.835167 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.835259 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.836415 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.836928 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.837170 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.837233 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.837644 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.837757 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.837952 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.837963 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.838465 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.839291 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.839771 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.839901 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.842138 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.842868 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.843167 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.843237 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.843285 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.843309 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.843659 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.844026 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.844507 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.844517 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.844605 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.844871 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.845040 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.845159 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.845247 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.845320 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.845517 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.845609 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.845840 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.846103 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.846262 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.846332 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.846343 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.846453 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.846659 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.846879 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.846895 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.847116 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.847236 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.847128 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.847682 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.847813 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.847873 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.847907 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848033 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848046 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848019 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848051 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848586 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848604 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848806 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848911 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848998 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.848949 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.849033 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.849225 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.849462 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.849523 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.849571 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.849538 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.849754 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.850294 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.850652 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.851047 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.850722 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.851177 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.851947 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.853721 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.854180 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.854267 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.855232 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.855286 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.855346 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.855790 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.855942 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.856140 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.855963 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.863424 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.865638 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.882919 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.883863 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.896640 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.898905 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678" exitCode=255 Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.900210 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678"} Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.910479 4948 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.910836 4948 scope.go:117] "RemoveContainer" containerID="65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911200 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911289 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911432 4948 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911452 4948 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911490 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911502 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911513 4948 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911524 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911534 4948 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911565 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911574 4948 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911583 4948 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911591 4948 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911600 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911609 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911617 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911652 4948 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911663 4948 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911677 4948 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911688 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911698 4948 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911727 4948 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911739 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911749 4948 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911758 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911768 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911779 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911805 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911816 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911827 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911841 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911825 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.911853 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912111 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912191 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912206 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912218 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912229 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912239 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912249 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912260 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912271 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912281 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912292 4948 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912303 4948 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912349 4948 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912361 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912372 4948 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912385 4948 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912774 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912819 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912829 4948 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912840 4948 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912850 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912859 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912870 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912880 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912889 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912899 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.912719 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913190 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913240 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913271 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913298 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913325 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913350 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913375 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913399 4948 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913423 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913447 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913473 4948 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913500 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913529 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913555 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913579 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913605 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913629 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913653 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913676 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913701 4948 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913725 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913752 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913776 4948 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913801 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913883 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913909 4948 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913935 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.913955 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914007 4948 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914033 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914059 4948 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914084 4948 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914108 4948 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914133 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914158 4948 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: E0220 08:06:12.914171 4948 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914183 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914209 4948 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914319 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914522 4948 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914539 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914549 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914559 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914569 4948 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914597 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914606 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914615 4948 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914625 4948 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914636 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914648 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914675 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914685 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914695 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914704 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914714 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914726 4948 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914759 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914771 4948 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914782 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914790 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914800 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914900 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914923 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914943 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.914963 4948 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915008 4948 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915026 4948 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915044 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915061 4948 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915077 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915094 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915110 4948 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915128 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915145 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915164 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915180 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915199 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915216 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915232 4948 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915249 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915265 4948 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915282 4948 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915299 4948 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915315 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915332 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915349 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915367 4948 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.915386 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.932726 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.948898 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.961296 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.964341 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.972066 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.977361 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.977723 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: I0220 08:06:12.989537 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:12 crc kubenswrapper[4948]: W0220 08:06:12.994417 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-db483d7683a8ed01afc8e5cfd09fa69c3b88dae193594965ab653e559b96f064 WatchSource:0}: Error finding container db483d7683a8ed01afc8e5cfd09fa69c3b88dae193594965ab653e559b96f064: Status 404 returned error can't find the container with id db483d7683a8ed01afc8e5cfd09fa69c3b88dae193594965ab653e559b96f064 Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.008334 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.018804 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.029415 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.037428 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-20 08:01:12 +0000 UTC, rotation deadline is 2027-01-01 04:33:54.676000435 +0000 UTC Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.037510 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7556h27m41.638495029s for next certificate rotation Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.319644 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.319735 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.319866 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.319859 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.319930 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:14.31991459 +0000 UTC m=+23.294409410 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.320020 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:14.319962711 +0000 UTC m=+23.294457561 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.420793 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.420861 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.420905 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421067 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421087 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421098 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421172 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:06:14.421103526 +0000 UTC m=+23.395598346 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421261 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:14.42124879 +0000 UTC m=+23.395743840 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421352 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421402 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421422 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:13 crc kubenswrapper[4948]: E0220 08:06:13.421523 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:14.421493596 +0000 UTC m=+23.395988456 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.664520 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 04:30:43.841625809 +0000 UTC Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.726923 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.727451 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.728396 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.729174 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.729870 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.730531 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.731261 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.731904 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.732771 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.735187 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.736008 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.737301 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.737908 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.739109 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.739761 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.740543 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.742350 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.743273 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.745400 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.746626 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.747637 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.749663 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.750581 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.752786 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.753650 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.755642 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.757571 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.758670 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.760094 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.761223 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.762288 4948 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.762508 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.768688 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.771426 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.772646 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.777259 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.780259 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.781890 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.783530 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.785260 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.786632 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.788404 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.790324 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.793119 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.794591 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.796917 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.799362 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.802151 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.803292 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.804599 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.805779 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.807291 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.808817 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.809596 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.843746 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.848696 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.875197 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.903673 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d"} Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.903802 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"93e398dada94802ff7db7fb1eb7f711e7a157be71ff2d442d8eb788b009b172a"} Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.905141 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.907083 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61"} Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.907208 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.909435 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b"} Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.909463 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237"} Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.909475 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"db483d7683a8ed01afc8e5cfd09fa69c3b88dae193594965ab653e559b96f064"} Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.910455 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1268deebbef00781e7b186ba2f25f8bfcbfb64830abd8c86bb416b2a6c52d395"} Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.928384 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:13Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.971622 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:13Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:13 crc kubenswrapper[4948]: I0220 08:06:13.988576 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:13Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.008366 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.026876 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.043572 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.060345 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.081770 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.094253 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-z8rrb"] Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.094561 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-z8rrb" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.102547 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.102735 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.103620 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.120341 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.147674 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.171161 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.185160 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.201474 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.227073 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.228420 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8c9cb414-e3c0-4e40-92b1-42f05565e44c-hosts-file\") pod \"node-resolver-z8rrb\" (UID: \"8c9cb414-e3c0-4e40-92b1-42f05565e44c\") " pod="openshift-dns/node-resolver-z8rrb" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.228476 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tzrg\" (UniqueName: \"kubernetes.io/projected/8c9cb414-e3c0-4e40-92b1-42f05565e44c-kube-api-access-4tzrg\") pod \"node-resolver-z8rrb\" (UID: \"8c9cb414-e3c0-4e40-92b1-42f05565e44c\") " pod="openshift-dns/node-resolver-z8rrb" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.242695 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.253513 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.264399 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.274036 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.286356 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.301524 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.316166 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.328904 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.328946 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8c9cb414-e3c0-4e40-92b1-42f05565e44c-hosts-file\") pod \"node-resolver-z8rrb\" (UID: \"8c9cb414-e3c0-4e40-92b1-42f05565e44c\") " pod="openshift-dns/node-resolver-z8rrb" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.328999 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.329025 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tzrg\" (UniqueName: \"kubernetes.io/projected/8c9cb414-e3c0-4e40-92b1-42f05565e44c-kube-api-access-4tzrg\") pod \"node-resolver-z8rrb\" (UID: \"8c9cb414-e3c0-4e40-92b1-42f05565e44c\") " pod="openshift-dns/node-resolver-z8rrb" Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.329120 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.329174 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8c9cb414-e3c0-4e40-92b1-42f05565e44c-hosts-file\") pod \"node-resolver-z8rrb\" (UID: \"8c9cb414-e3c0-4e40-92b1-42f05565e44c\") " pod="openshift-dns/node-resolver-z8rrb" Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.329209 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.329234 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:16.329209036 +0000 UTC m=+25.303703926 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.329328 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:16.329290518 +0000 UTC m=+25.303785338 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.332054 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.347788 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.359589 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.377498 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.397650 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.422688 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.429903 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.430257 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tzrg\" (UniqueName: \"kubernetes.io/projected/8c9cb414-e3c0-4e40-92b1-42f05565e44c-kube-api-access-4tzrg\") pod \"node-resolver-z8rrb\" (UID: \"8c9cb414-e3c0-4e40-92b1-42f05565e44c\") " pod="openshift-dns/node-resolver-z8rrb" Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.430263 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:06:16.430226737 +0000 UTC m=+25.404721557 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.430494 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.430618 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.430722 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.430752 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.430770 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.430822 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:16.430804062 +0000 UTC m=+25.405298882 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.431055 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.431148 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.431215 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.431353 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:16.431326405 +0000 UTC m=+25.405821465 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.665087 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 05:20:23.981600229 +0000 UTC Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.705303 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-z8rrb" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.724197 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.724234 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.724290 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.724496 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.724333 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:14 crc kubenswrapper[4948]: E0220 08:06:14.724695 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.877942 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-8frp4"] Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.878575 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.879130 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-h4ww2"] Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.879363 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.879424 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-x7s56"] Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.879890 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-x7s56" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.880151 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.881089 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.881597 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.881658 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.882897 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.883336 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.883830 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.884222 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.884259 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.884591 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.884856 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.885413 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.912440 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.920383 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-z8rrb" event={"ID":"8c9cb414-e3c0-4e40-92b1-42f05565e44c","Type":"ContainerStarted","Data":"0482621a1c61d8c484fe2f3b29c02bacf568c4be68044be2ed544076db12509f"} Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.933554 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.957092 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.969047 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:14 crc kubenswrapper[4948]: I0220 08:06:14.994325 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.007987 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.022663 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.035907 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5757f611-4b0c-49dc-a076-c90ebaad78d8-cni-binary-copy\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.035965 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-mcd-auth-proxy-config\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036009 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b96124e4-0a74-4578-9142-fd728eb9f99e-cni-binary-copy\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036035 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-os-release\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036108 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-kubelet\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036159 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-daemon-config\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036175 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-etc-kubernetes\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036199 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-os-release\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036218 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-cnibin\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036234 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5757f611-4b0c-49dc-a076-c90ebaad78d8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036260 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-cnibin\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036300 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036364 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-cni-multus\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036428 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-hostroot\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036449 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-multus-certs\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036473 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9689\" (UniqueName: \"kubernetes.io/projected/5757f611-4b0c-49dc-a076-c90ebaad78d8-kube-api-access-x9689\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036493 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnmlr\" (UniqueName: \"kubernetes.io/projected/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-kube-api-access-gnmlr\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036508 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-system-cni-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036524 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-proxy-tls\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036540 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-cni-bin\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036576 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-socket-dir-parent\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036593 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-cni-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036606 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2fpv\" (UniqueName: \"kubernetes.io/projected/b96124e4-0a74-4578-9142-fd728eb9f99e-kube-api-access-b2fpv\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036624 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-conf-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036661 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-rootfs\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036750 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-k8s-cni-cncf-io\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036815 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-system-cni-dir\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.036842 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-netns\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.042760 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.069929 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.099655 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.118546 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.137808 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-rootfs\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.137864 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-k8s-cni-cncf-io\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.137896 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-system-cni-dir\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.137918 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-netns\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.137940 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5757f611-4b0c-49dc-a076-c90ebaad78d8-cni-binary-copy\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.137961 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b96124e4-0a74-4578-9142-fd728eb9f99e-cni-binary-copy\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138001 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-mcd-auth-proxy-config\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.137996 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-rootfs\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138025 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-k8s-cni-cncf-io\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138082 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-netns\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138052 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-kubelet\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138000 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-system-cni-dir\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138024 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-kubelet\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138150 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-daemon-config\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138173 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-etc-kubernetes\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138188 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-os-release\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138207 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-os-release\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138224 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-cnibin\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138242 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-cnibin\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138259 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5757f611-4b0c-49dc-a076-c90ebaad78d8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138294 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138315 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9689\" (UniqueName: \"kubernetes.io/projected/5757f611-4b0c-49dc-a076-c90ebaad78d8-kube-api-access-x9689\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138337 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-cni-multus\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138353 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-hostroot\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138366 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-multus-certs\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138385 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnmlr\" (UniqueName: \"kubernetes.io/projected/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-kube-api-access-gnmlr\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138408 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-system-cni-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138427 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-cni-bin\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138442 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-proxy-tls\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138466 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-socket-dir-parent\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138483 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-cni-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138496 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2fpv\" (UniqueName: \"kubernetes.io/projected/b96124e4-0a74-4578-9142-fd728eb9f99e-kube-api-access-b2fpv\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138511 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-conf-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138561 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-conf-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138668 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b96124e4-0a74-4578-9142-fd728eb9f99e-cni-binary-copy\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138697 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5757f611-4b0c-49dc-a076-c90ebaad78d8-cni-binary-copy\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138724 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-cni-multus\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138742 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-var-lib-cni-bin\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138751 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-etc-kubernetes\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138768 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-hostroot\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138794 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-host-run-multus-certs\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138837 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-mcd-auth-proxy-config\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.138911 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-cnibin\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139065 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-os-release\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139083 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-system-cni-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139122 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-cni-dir\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139127 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-os-release\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139159 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-socket-dir-parent\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139178 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b96124e4-0a74-4578-9142-fd728eb9f99e-cnibin\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139242 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b96124e4-0a74-4578-9142-fd728eb9f99e-multus-daemon-config\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139558 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5757f611-4b0c-49dc-a076-c90ebaad78d8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.139691 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5757f611-4b0c-49dc-a076-c90ebaad78d8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.140377 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.143373 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-proxy-tls\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.160336 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2fpv\" (UniqueName: \"kubernetes.io/projected/b96124e4-0a74-4578-9142-fd728eb9f99e-kube-api-access-b2fpv\") pod \"multus-x7s56\" (UID: \"b96124e4-0a74-4578-9142-fd728eb9f99e\") " pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.168710 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnmlr\" (UniqueName: \"kubernetes.io/projected/84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee-kube-api-access-gnmlr\") pod \"machine-config-daemon-h4ww2\" (UID: \"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\") " pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.171854 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.180543 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9689\" (UniqueName: \"kubernetes.io/projected/5757f611-4b0c-49dc-a076-c90ebaad78d8-kube-api-access-x9689\") pod \"multus-additional-cni-plugins-8frp4\" (UID: \"5757f611-4b0c-49dc-a076-c90ebaad78d8\") " pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.190265 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.191287 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-8frp4" Feb 20 08:06:15 crc kubenswrapper[4948]: W0220 08:06:15.202365 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5757f611_4b0c_49dc_a076_c90ebaad78d8.slice/crio-ea4516e20b5daad40c01344cd4ecddf48b4f75302941031c96ccc00b8502043a WatchSource:0}: Error finding container ea4516e20b5daad40c01344cd4ecddf48b4f75302941031c96ccc00b8502043a: Status 404 returned error can't find the container with id ea4516e20b5daad40c01344cd4ecddf48b4f75302941031c96ccc00b8502043a Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.202349 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.215016 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.221238 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.227179 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-x7s56" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.235922 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: W0220 08:06:15.245109 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb96124e4_0a74_4578_9142_fd728eb9f99e.slice/crio-cb3225162882e7895b01638438f3868817f5fc21d94397971005435bdd093c67 WatchSource:0}: Error finding container cb3225162882e7895b01638438f3868817f5fc21d94397971005435bdd093c67: Status 404 returned error can't find the container with id cb3225162882e7895b01638438f3868817f5fc21d94397971005435bdd093c67 Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.252866 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.266433 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7s8gg"] Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.267354 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.271069 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.271660 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.271736 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.271998 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.272512 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.272836 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.278880 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.278998 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.292993 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.306844 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.317327 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.332726 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.340884 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-ovn-kubernetes\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.340912 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-slash\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.340927 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-var-lib-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.340951 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-etc-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.340964 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-netd\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.340990 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341025 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovn-node-metrics-cert\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341079 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-systemd\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341121 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-log-socket\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341149 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtrs5\" (UniqueName: \"kubernetes.io/projected/d6a28bd7-2b83-43f8-b803-bfe41516e071-kube-api-access-gtrs5\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341186 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-kubelet\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341204 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341227 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-node-log\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341246 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-env-overrides\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341266 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-systemd-units\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341285 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-config\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341310 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-bin\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341337 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-script-lib\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341354 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-netns\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.341388 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-ovn\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.349893 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.371516 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.387606 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.400215 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.414053 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.425998 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442536 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-etc-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442580 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-netd\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442598 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442671 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovn-node-metrics-cert\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442693 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-systemd\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442709 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtrs5\" (UniqueName: \"kubernetes.io/projected/d6a28bd7-2b83-43f8-b803-bfe41516e071-kube-api-access-gtrs5\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442734 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-log-socket\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442771 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-kubelet\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442790 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442811 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-node-log\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442827 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-env-overrides\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442847 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-systemd-units\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442863 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-config\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442881 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-bin\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442897 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-script-lib\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442922 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-netns\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442942 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-ovn\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.442960 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-ovn-kubernetes\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443001 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-slash\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443019 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-var-lib-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443091 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-var-lib-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443132 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-etc-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443156 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-netd\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443178 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443203 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-systemd-units\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443225 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-ovn\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443248 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-ovn-kubernetes\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443301 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-netns\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443359 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-systemd\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443377 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-openvswitch\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443387 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-kubelet\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443391 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-bin\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443406 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-node-log\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443393 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-slash\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443540 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-log-socket\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.443907 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-env-overrides\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.444203 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-script-lib\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.444355 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-config\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.448921 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.449143 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovn-node-metrics-cert\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.464556 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtrs5\" (UniqueName: \"kubernetes.io/projected/d6a28bd7-2b83-43f8-b803-bfe41516e071-kube-api-access-gtrs5\") pod \"ovnkube-node-7s8gg\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.465828 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.482547 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.498928 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.512247 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.525712 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.538301 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.550503 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.562931 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.665376 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 10:31:20.104008261 +0000 UTC Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.673733 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:15 crc kubenswrapper[4948]: W0220 08:06:15.685485 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6a28bd7_2b83_43f8_b803_bfe41516e071.slice/crio-0269d0e4618f9bfc08bbd4b396cc89a6bd78be17afe5c9425cf43f02da885bc5 WatchSource:0}: Error finding container 0269d0e4618f9bfc08bbd4b396cc89a6bd78be17afe5c9425cf43f02da885bc5: Status 404 returned error can't find the container with id 0269d0e4618f9bfc08bbd4b396cc89a6bd78be17afe5c9425cf43f02da885bc5 Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.926443 4948 generic.go:334] "Generic (PLEG): container finished" podID="5757f611-4b0c-49dc-a076-c90ebaad78d8" containerID="dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069" exitCode=0 Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.926522 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" event={"ID":"5757f611-4b0c-49dc-a076-c90ebaad78d8","Type":"ContainerDied","Data":"dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.926559 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" event={"ID":"5757f611-4b0c-49dc-a076-c90ebaad78d8","Type":"ContainerStarted","Data":"ea4516e20b5daad40c01344cd4ecddf48b4f75302941031c96ccc00b8502043a"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.928320 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7s56" event={"ID":"b96124e4-0a74-4578-9142-fd728eb9f99e","Type":"ContainerStarted","Data":"34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.928350 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7s56" event={"ID":"b96124e4-0a74-4578-9142-fd728eb9f99e","Type":"ContainerStarted","Data":"cb3225162882e7895b01638438f3868817f5fc21d94397971005435bdd093c67"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.934749 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.937770 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad" exitCode=0 Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.937860 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.938036 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"0269d0e4618f9bfc08bbd4b396cc89a6bd78be17afe5c9425cf43f02da885bc5"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.940180 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.940206 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.940217 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"d824cc19743acd2484c541a11dd55f5e646b98af34783b9779359153056f8cb3"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.942217 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-z8rrb" event={"ID":"8c9cb414-e3c0-4e40-92b1-42f05565e44c","Type":"ContainerStarted","Data":"180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b"} Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.960422 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:15 crc kubenswrapper[4948]: I0220 08:06:15.977260 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.007516 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.024608 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.037540 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.053395 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.068282 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.087690 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.102849 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.120560 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.134222 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.145138 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.180893 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.197952 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.224886 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.236691 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.251304 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.265016 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.288395 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.301391 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.322414 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.337587 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.352037 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.354528 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.354577 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.354702 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.354708 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.354771 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:20.354752592 +0000 UTC m=+29.329247412 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.354809 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:20.354783963 +0000 UTC m=+29.329278793 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.366090 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.386930 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.401678 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.417164 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.433334 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.455736 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.455887 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:06:20.455860779 +0000 UTC m=+29.430355599 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.455946 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.456101 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.456131 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.456144 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.456137 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.456181 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:20.456171786 +0000 UTC m=+29.430666736 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.456309 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.456343 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.456358 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.456418 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:20.456399372 +0000 UTC m=+29.430894242 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.666043 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 12:39:18.947377048 +0000 UTC Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.721612 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.721671 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.721735 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.721774 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.721957 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:16 crc kubenswrapper[4948]: E0220 08:06:16.722179 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.950727 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.951055 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.951072 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.951087 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.951100 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.953015 4948 generic.go:334] "Generic (PLEG): container finished" podID="5757f611-4b0c-49dc-a076-c90ebaad78d8" containerID="9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32" exitCode=0 Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.953102 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" event={"ID":"5757f611-4b0c-49dc-a076-c90ebaad78d8","Type":"ContainerDied","Data":"9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32"} Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.976173 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:16 crc kubenswrapper[4948]: I0220 08:06:16.997375 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:16Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.014738 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.043436 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.069395 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.085716 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.107666 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.132203 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.157909 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.158390 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-nsk7l"] Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.158862 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.160715 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.161037 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.161175 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.161475 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.171761 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.189637 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.205573 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.219902 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.236185 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.261330 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.267617 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc562\" (UniqueName: \"kubernetes.io/projected/ba900a81-fc7b-40b3-942c-097c47812bf4-kube-api-access-dc562\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.267756 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ba900a81-fc7b-40b3-942c-097c47812bf4-serviceca\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.267815 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ba900a81-fc7b-40b3-942c-097c47812bf4-host\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.276299 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.294285 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.307596 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.334081 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.349270 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.362547 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.369177 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ba900a81-fc7b-40b3-942c-097c47812bf4-host\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.369347 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ba900a81-fc7b-40b3-942c-097c47812bf4-host\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.369270 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc562\" (UniqueName: \"kubernetes.io/projected/ba900a81-fc7b-40b3-942c-097c47812bf4-kube-api-access-dc562\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.369484 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ba900a81-fc7b-40b3-942c-097c47812bf4-serviceca\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.370601 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ba900a81-fc7b-40b3-942c-097c47812bf4-serviceca\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.385883 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.399745 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc562\" (UniqueName: \"kubernetes.io/projected/ba900a81-fc7b-40b3-942c-097c47812bf4-kube-api-access-dc562\") pod \"node-ca-nsk7l\" (UID: \"ba900a81-fc7b-40b3-942c-097c47812bf4\") " pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.405132 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.416760 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.432459 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.445766 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.460792 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.471647 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-nsk7l" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.471700 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.484777 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.666747 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 11:35:45.396778704 +0000 UTC Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.960717 4948 generic.go:334] "Generic (PLEG): container finished" podID="5757f611-4b0c-49dc-a076-c90ebaad78d8" containerID="c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6" exitCode=0 Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.960799 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" event={"ID":"5757f611-4b0c-49dc-a076-c90ebaad78d8","Type":"ContainerDied","Data":"c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6"} Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.966229 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.967744 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-nsk7l" event={"ID":"ba900a81-fc7b-40b3-942c-097c47812bf4","Type":"ContainerStarted","Data":"eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7"} Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.967796 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-nsk7l" event={"ID":"ba900a81-fc7b-40b3-942c-097c47812bf4","Type":"ContainerStarted","Data":"22fe034e4c89f89b70792ebcca37ec652f2177605c3edc3505d40246c825794e"} Feb 20 08:06:17 crc kubenswrapper[4948]: I0220 08:06:17.987603 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.007311 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.037632 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.060671 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.076490 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.099027 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.121767 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.141719 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.162436 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.186086 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.201655 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.219537 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.248627 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.266697 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.282950 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.304657 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.318936 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.337554 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.356769 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.388820 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.391611 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.391667 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.391684 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.391827 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.403205 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.406877 4948 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.407305 4948 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.408736 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.408779 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.408798 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.408820 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.408837 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.427608 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.443200 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.448871 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.456057 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.456101 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.456113 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.456131 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.456143 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.463657 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.469232 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.472509 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.472584 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.472601 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.472626 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.472642 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.483052 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.486737 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.491107 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.491162 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.491174 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.491196 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.491224 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.501369 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.504525 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.509597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.509638 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.509686 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.509710 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.509725 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.510940 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.520571 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.520692 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.522304 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.522331 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.522342 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.522360 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.522373 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.551704 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.590946 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.626065 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.626113 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.626125 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.626142 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.626154 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.627499 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.664957 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.667306 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 00:39:35.321571782 +0000 UTC Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.722280 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.722408 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.722527 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.722575 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.722765 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:18 crc kubenswrapper[4948]: E0220 08:06:18.722870 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.729657 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.729715 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.729733 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.729786 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.729806 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.833425 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.833482 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.833499 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.833521 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.833540 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.943034 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.943096 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.943114 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.943139 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.943158 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:18Z","lastTransitionTime":"2026-02-20T08:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.977052 4948 generic.go:334] "Generic (PLEG): container finished" podID="5757f611-4b0c-49dc-a076-c90ebaad78d8" containerID="b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155" exitCode=0 Feb 20 08:06:18 crc kubenswrapper[4948]: I0220 08:06:18.977111 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" event={"ID":"5757f611-4b0c-49dc-a076-c90ebaad78d8","Type":"ContainerDied","Data":"b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.000573 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:18Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.026153 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.047298 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.047348 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.047366 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.047391 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.047410 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.051518 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.077045 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.095014 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.112184 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.129938 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.157668 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.157735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.157748 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.157774 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.157791 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.162243 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.185260 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.198319 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.227427 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.244967 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.260761 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.260802 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.260812 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.260830 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.260845 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.271459 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.289540 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.305821 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:19Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.363576 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.363620 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.363632 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.363652 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.363666 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.467182 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.467257 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.467288 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.467321 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.467340 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.570952 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.571032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.571045 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.571066 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.571078 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.667834 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 11:54:04.715149774 +0000 UTC Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.675325 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.675371 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.675384 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.675404 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.675419 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.779109 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.779190 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.779209 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.779237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.779256 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.882199 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.882271 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.882291 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.882319 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.882341 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.990603 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.990660 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.990681 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.990708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.990734 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:19Z","lastTransitionTime":"2026-02-20T08:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.995261 4948 generic.go:334] "Generic (PLEG): container finished" podID="5757f611-4b0c-49dc-a076-c90ebaad78d8" containerID="1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc" exitCode=0 Feb 20 08:06:19 crc kubenswrapper[4948]: I0220 08:06:19.995345 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" event={"ID":"5757f611-4b0c-49dc-a076-c90ebaad78d8","Type":"ContainerDied","Data":"1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.006362 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.035293 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.060217 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.080300 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.093924 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.094012 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.094031 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.094058 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.094079 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.101554 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.131324 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.156412 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.175189 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.199053 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.199154 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.199220 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.199251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.199317 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.205483 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.227358 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.253321 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.274269 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.298487 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.302909 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.302997 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.303010 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.303039 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.303055 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.319478 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.337960 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.359279 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.403684 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.403763 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.403947 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.404078 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:28.404054763 +0000 UTC m=+37.378549603 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.404647 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.404727 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:28.404708739 +0000 UTC m=+37.379203569 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.407112 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.407153 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.407166 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.407185 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.407199 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.504894 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505177 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:06:28.505139375 +0000 UTC m=+37.479634235 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.505334 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.505453 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505617 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505645 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505660 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505674 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505688 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505695 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505760 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:28.505739869 +0000 UTC m=+37.480234719 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.505786 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:28.50577459 +0000 UTC m=+37.480269450 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.510613 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.510670 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.510686 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.510712 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.510732 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.614001 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.614072 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.614088 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.614111 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.614128 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.668359 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 03:45:56.265401649 +0000 UTC Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.716622 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.716676 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.716691 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.716708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.716722 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.722074 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.722074 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.722195 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.722271 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.722093 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:20 crc kubenswrapper[4948]: E0220 08:06:20.722341 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.820163 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.820208 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.820220 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.820237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.820249 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.922820 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.922916 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.922939 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.922968 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:20 crc kubenswrapper[4948]: I0220 08:06:20.923026 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:20Z","lastTransitionTime":"2026-02-20T08:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.014658 4948 generic.go:334] "Generic (PLEG): container finished" podID="5757f611-4b0c-49dc-a076-c90ebaad78d8" containerID="86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4" exitCode=0 Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.014730 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" event={"ID":"5757f611-4b0c-49dc-a076-c90ebaad78d8","Type":"ContainerDied","Data":"86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.025452 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.025516 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.025537 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.025560 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.025578 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.038768 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.061453 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.075459 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.094900 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.117133 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.130614 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.130687 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.130712 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.130743 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.130766 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.134007 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.148123 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.167773 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.182035 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.198393 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.210379 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.234185 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.234237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.234249 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.234268 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.234283 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.235848 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.256884 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.274656 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.291268 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.337645 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.337700 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.337718 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.337745 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.337768 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.440349 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.440439 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.440507 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.440541 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.440564 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.487160 4948 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.543469 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.543658 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.543759 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.544029 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.544113 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.647427 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.647496 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.647525 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.647559 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.647583 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.669123 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 23:46:06.816799505 +0000 UTC Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.750756 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.750818 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.750837 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.750862 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.750881 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.752258 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.775145 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.795726 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.828038 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.854747 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.854818 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.854842 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.854874 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.854898 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.858440 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.880641 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.901628 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.928413 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.952616 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.958060 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.958105 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.958124 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.958150 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.958171 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:21Z","lastTransitionTime":"2026-02-20T08:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.967659 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:21 crc kubenswrapper[4948]: I0220 08:06:21.997967 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.017086 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.023504 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" event={"ID":"5757f611-4b0c-49dc-a076-c90ebaad78d8","Type":"ContainerStarted","Data":"78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.031323 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.032199 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.036743 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.061367 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.061431 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.061447 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.061474 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.061494 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.063557 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.081246 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.085416 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.098141 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.116582 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.134227 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.165307 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.165356 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.165373 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.165396 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.165416 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.169452 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.188392 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.207287 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.228263 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.249531 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.269393 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.269460 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.269485 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.269515 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.269539 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.270569 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.289762 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.311813 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.329496 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.347210 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.373373 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.373416 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.373445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.373464 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.373475 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.384885 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.405298 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.424657 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.441220 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.455109 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.470131 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.476160 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.476229 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.476239 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.476257 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.476269 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.504192 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.522546 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.537310 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.557341 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.573532 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.579045 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.579095 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.579114 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.579140 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.579159 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.592562 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.607717 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.637584 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.659343 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.669708 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 07:19:09.782432833 +0000 UTC Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.676324 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.682078 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.682153 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.682183 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.682210 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.682227 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.693842 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.722099 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.722099 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:22 crc kubenswrapper[4948]: E0220 08:06:22.722261 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:22 crc kubenswrapper[4948]: E0220 08:06:22.722422 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.722723 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:22 crc kubenswrapper[4948]: E0220 08:06:22.723041 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.784437 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.784716 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.784802 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.784884 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.784990 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.888370 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.888731 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.888941 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.889187 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.889323 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.993040 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.993105 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.993128 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.993160 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:22 crc kubenswrapper[4948]: I0220 08:06:22.993182 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:22Z","lastTransitionTime":"2026-02-20T08:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.034897 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.035627 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.070388 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.094643 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.097075 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.097142 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.097162 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.097193 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.097210 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.111167 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.136507 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.155937 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.188164 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.200424 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.200861 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.200881 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.200906 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.200927 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.211826 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.232669 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.253046 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.273850 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.300040 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.304366 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.304436 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.304461 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.304491 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.304517 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.327025 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.358376 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.393903 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.407619 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.407697 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.407722 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.407758 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.407781 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.417730 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.435316 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:23Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.510598 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.510663 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.510685 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.510710 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.510728 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.613550 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.613614 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.613632 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.613663 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.613682 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.670876 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 10:33:34.186661742 +0000 UTC Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.716400 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.716464 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.716482 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.716506 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.716527 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.820526 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.820579 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.820590 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.820608 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.820619 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.923684 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.923735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.923746 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.923765 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:23 crc kubenswrapper[4948]: I0220 08:06:23.923777 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:23Z","lastTransitionTime":"2026-02-20T08:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.026501 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.026559 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.026579 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.026605 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.026623 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.037737 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.129487 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.129555 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.129573 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.129599 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.129617 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.232433 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.232507 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.232522 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.232541 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.232580 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.336036 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.336071 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.336080 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.336095 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.336105 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.443880 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.444218 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.444328 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.445350 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.445441 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.548706 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.548767 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.548785 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.548814 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.548838 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.652039 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.652129 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.652149 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.652175 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.652194 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.672853 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 05:25:01.737493783 +0000 UTC Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.722366 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.722440 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.722436 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:24 crc kubenswrapper[4948]: E0220 08:06:24.722569 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:24 crc kubenswrapper[4948]: E0220 08:06:24.722655 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:24 crc kubenswrapper[4948]: E0220 08:06:24.722761 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.755156 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.755251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.755296 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.755325 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.755340 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.858319 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.858402 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.858425 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.858454 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.858474 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.961144 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.961216 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.961235 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.961261 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:24 crc kubenswrapper[4948]: I0220 08:06:24.961280 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:24Z","lastTransitionTime":"2026-02-20T08:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.045181 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/0.log" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.048935 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a" exitCode=1 Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.049045 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.050191 4948 scope.go:117] "RemoveContainer" containerID="3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.064671 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.064758 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.064787 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.064819 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.064845 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.089598 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.114878 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.131083 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.155845 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.167498 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.167555 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.167575 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.167600 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.167619 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.174367 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.209836 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:24Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:06:24.465835 6302 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:24.466938 6302 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 08:06:24.467064 6302 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:24.467092 6302 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 08:06:24.467102 6302 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 08:06:24.467169 6302 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:24.467179 6302 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:24.467254 6302 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 08:06:24.467267 6302 factory.go:656] Stopping watch factory\\\\nI0220 08:06:24.467295 6302 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:24.467300 6302 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 08:06:24.467312 6302 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:24.467326 6302 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 08:06:24.467329 6302 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.237349 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.257878 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.270548 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.270602 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.270622 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.270666 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.270686 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.279617 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.303008 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.321387 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.344358 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.362098 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.373903 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.374016 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.374037 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.374063 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.374548 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.381408 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.402025 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:25Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.478284 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.478330 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.478342 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.478362 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.478376 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.581268 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.581317 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.581326 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.581344 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.581355 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.673015 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 06:41:28.43946064 +0000 UTC Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.684471 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.684536 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.684561 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.684596 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.684622 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.788003 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.788063 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.788083 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.788106 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.788130 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.891158 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.891227 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.891242 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.891268 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.891282 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.995201 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.995271 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.995289 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.995313 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:25 crc kubenswrapper[4948]: I0220 08:06:25.995332 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:25Z","lastTransitionTime":"2026-02-20T08:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.058001 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/0.log" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.062769 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.062894 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.098643 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.098855 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.099033 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.099195 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.099317 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.104181 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.126087 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.143220 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.164061 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.179458 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.201844 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.202558 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.202624 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.202643 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.202668 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.202688 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.220381 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.252145 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:24Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:06:24.465835 6302 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:24.466938 6302 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 08:06:24.467064 6302 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:24.467092 6302 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 08:06:24.467102 6302 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 08:06:24.467169 6302 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:24.467179 6302 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:24.467254 6302 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 08:06:24.467267 6302 factory.go:656] Stopping watch factory\\\\nI0220 08:06:24.467295 6302 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:24.467300 6302 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 08:06:24.467312 6302 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:24.467326 6302 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 08:06:24.467329 6302 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.276108 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.299492 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.305423 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.305476 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.305492 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.305518 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.305536 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.321097 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.347915 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.370409 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.390278 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.408390 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.408458 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.408482 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.408513 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.408541 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.413096 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:26Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.511397 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.511439 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.511450 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.511466 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.511478 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.615021 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.615098 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.615123 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.615154 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.615180 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.674292 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 12:05:26.593601973 +0000 UTC Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.718235 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.718328 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.718348 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.718374 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.718393 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.721611 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.721640 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.721610 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:26 crc kubenswrapper[4948]: E0220 08:06:26.721772 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:26 crc kubenswrapper[4948]: E0220 08:06:26.721921 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:26 crc kubenswrapper[4948]: E0220 08:06:26.722134 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.821799 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.821928 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.821958 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.822036 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.822063 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.924954 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.925387 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.925521 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.925622 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:26 crc kubenswrapper[4948]: I0220 08:06:26.925730 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:26Z","lastTransitionTime":"2026-02-20T08:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.028839 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.028896 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.028913 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.028938 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.028960 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.067861 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/1.log" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.068444 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/0.log" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.071775 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e" exitCode=1 Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.071831 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.071898 4948 scope.go:117] "RemoveContainer" containerID="3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.072863 4948 scope.go:117] "RemoveContainer" containerID="8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e" Feb 20 08:06:27 crc kubenswrapper[4948]: E0220 08:06:27.073135 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.091033 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.107040 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.130622 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.132120 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.132170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.132188 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.132213 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.132237 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.149854 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.180368 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:24Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:06:24.465835 6302 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:24.466938 6302 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 08:06:24.467064 6302 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:24.467092 6302 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 08:06:24.467102 6302 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 08:06:24.467169 6302 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:24.467179 6302 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:24.467254 6302 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 08:06:24.467267 6302 factory.go:656] Stopping watch factory\\\\nI0220 08:06:24.467295 6302 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:24.467300 6302 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 08:06:24.467312 6302 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:24.467326 6302 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 08:06:24.467329 6302 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.202770 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.224061 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.235507 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.235578 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.235600 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.235628 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.235647 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.245288 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.266103 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.288393 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.301331 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.308604 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.327325 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.341026 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.341107 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.341127 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.341568 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.341622 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.358906 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.377757 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.392455 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.414934 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.430555 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.445322 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.445407 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.445432 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.445466 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.445491 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.469058 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.491672 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.507686 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.531366 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.548955 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.549058 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.549078 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.549105 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.549129 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.551906 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.587080 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:24Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:06:24.465835 6302 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:24.466938 6302 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 08:06:24.467064 6302 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:24.467092 6302 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 08:06:24.467102 6302 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 08:06:24.467169 6302 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:24.467179 6302 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:24.467254 6302 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 08:06:24.467267 6302 factory.go:656] Stopping watch factory\\\\nI0220 08:06:24.467295 6302 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:24.467300 6302 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 08:06:24.467312 6302 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:24.467326 6302 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 08:06:24.467329 6302 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.608751 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.629829 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.652619 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.653038 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.653068 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.653097 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.653118 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.652604 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.675053 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 03:55:57.781471648 +0000 UTC Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.677741 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.698557 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.722422 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.742402 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:27Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.756900 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.756952 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.756996 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.757021 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.757040 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.860602 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.860711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.860770 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.860801 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.860820 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.964738 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.964825 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.964859 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.964886 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:27 crc kubenswrapper[4948]: I0220 08:06:27.964908 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:27Z","lastTransitionTime":"2026-02-20T08:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.002023 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql"] Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.002693 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.006535 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.006628 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.033797 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.058519 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.067907 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.068019 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.068038 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.068063 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.068083 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.080659 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.083826 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/1.log" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.086498 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhj9x\" (UniqueName: \"kubernetes.io/projected/1b664cac-f1be-4b54-b85d-fede8b4d7d31-kube-api-access-qhj9x\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.086574 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1b664cac-f1be-4b54-b85d-fede8b4d7d31-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.086850 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1b664cac-f1be-4b54-b85d-fede8b4d7d31-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.086956 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1b664cac-f1be-4b54-b85d-fede8b4d7d31-env-overrides\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.091648 4948 scope.go:117] "RemoveContainer" containerID="8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.092122 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.100427 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.123205 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.144945 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.163708 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.171727 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.171807 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.171835 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.171874 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.171898 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.184481 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.187917 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1b664cac-f1be-4b54-b85d-fede8b4d7d31-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.188032 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1b664cac-f1be-4b54-b85d-fede8b4d7d31-env-overrides\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.188184 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhj9x\" (UniqueName: \"kubernetes.io/projected/1b664cac-f1be-4b54-b85d-fede8b4d7d31-kube-api-access-qhj9x\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.188236 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1b664cac-f1be-4b54-b85d-fede8b4d7d31-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.189491 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1b664cac-f1be-4b54-b85d-fede8b4d7d31-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.190179 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1b664cac-f1be-4b54-b85d-fede8b4d7d31-env-overrides\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.197910 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1b664cac-f1be-4b54-b85d-fede8b4d7d31-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.209355 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhj9x\" (UniqueName: \"kubernetes.io/projected/1b664cac-f1be-4b54-b85d-fede8b4d7d31-kube-api-access-qhj9x\") pod \"ovnkube-control-plane-749d76644c-j8xql\" (UID: \"1b664cac-f1be-4b54-b85d-fede8b4d7d31\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.222033 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.244669 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.262746 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.274796 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.274871 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.274896 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.274927 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.274953 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.286743 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.306089 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.324898 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.343512 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3456b89afa02fb3c4b83afea773783e0ae1c65903d24789d1a742688b7745b1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:24Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:06:24.465835 6302 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:24.466938 6302 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 08:06:24.467064 6302 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:24.467092 6302 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 08:06:24.467102 6302 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 08:06:24.467169 6302 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:24.467179 6302 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:24.467254 6302 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 08:06:24.467267 6302 factory.go:656] Stopping watch factory\\\\nI0220 08:06:24.467295 6302 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:24.467300 6302 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 08:06:24.467312 6302 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:24.467326 6302 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 08:06:24.467329 6302 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: W0220 08:06:28.348509 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b664cac_f1be_4b54_b85d_fede8b4d7d31.slice/crio-d7050c82a293de4981bb1136cb34984f9983b39afd30346c219c09e99e0451c4 WatchSource:0}: Error finding container d7050c82a293de4981bb1136cb34984f9983b39afd30346c219c09e99e0451c4: Status 404 returned error can't find the container with id d7050c82a293de4981bb1136cb34984f9983b39afd30346c219c09e99e0451c4 Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.369151 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.377886 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.377933 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.377949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.377995 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.378009 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.389795 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.408437 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.424370 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.460314 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.477688 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.481178 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.481278 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.481305 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.481345 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.481377 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.490441 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.490501 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.490659 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.490668 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.490746 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:44.490725011 +0000 UTC m=+53.465219861 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.490790 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:44.490762622 +0000 UTC m=+53.465257642 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.494543 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.519469 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.538653 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.565451 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.585492 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.585571 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.585591 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.585618 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.585638 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.587135 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.591321 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.591485 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:06:44.591456824 +0000 UTC m=+53.565951654 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.591532 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.591621 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.591740 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.591758 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.591774 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.591825 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:44.591814382 +0000 UTC m=+53.566309212 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.591883 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.591919 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.591943 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.592085 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 08:06:44.592050468 +0000 UTC m=+53.566545348 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.608418 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.625147 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.651792 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.676303 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 19:08:58.852604819 +0000 UTC Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.677336 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.693579 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.693639 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.693727 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.693790 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.693811 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.705024 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.721496 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.721539 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.721623 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.721671 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.721804 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.722003 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.722092 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.740836 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.770104 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-xd86g"] Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.770868 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.771007 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.793581 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.793673 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdbl2\" (UniqueName: \"kubernetes.io/projected/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-kube-api-access-gdbl2\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.797374 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.797426 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.797440 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.797461 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.797476 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.821055 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.843020 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.850176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.850218 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.850228 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.850247 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.850259 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.858174 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.866174 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.870019 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.870058 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.870069 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.870086 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.870117 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.875243 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.882842 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.886512 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.887772 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.887825 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.887834 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.887852 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.887863 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.894325 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.894362 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdbl2\" (UniqueName: \"kubernetes.io/projected/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-kube-api-access-gdbl2\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.894474 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.894530 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs podName:62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd nodeName:}" failed. No retries permitted until 2026-02-20 08:06:29.394516213 +0000 UTC m=+38.369011033 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs") pod "network-metrics-daemon-xd86g" (UID: "62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.900550 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.902464 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.905856 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.905887 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.905894 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.905909 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.905919 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.915910 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdbl2\" (UniqueName: \"kubernetes.io/projected/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-kube-api-access-gdbl2\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.917716 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.920113 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.924666 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.924718 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.924731 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.924751 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.924764 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.932456 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.938471 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: E0220 08:06:28.938575 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.940139 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.940165 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.940176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.940191 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.940202 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:28Z","lastTransitionTime":"2026-02-20T08:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.947747 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.968117 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:28 crc kubenswrapper[4948]: I0220 08:06:28.986896 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:28Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.005385 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.030886 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.042733 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.042787 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.042800 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.042818 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.042831 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.047738 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.061351 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.075788 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.097596 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.100732 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" event={"ID":"1b664cac-f1be-4b54-b85d-fede8b4d7d31","Type":"ContainerStarted","Data":"b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.100854 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" event={"ID":"1b664cac-f1be-4b54-b85d-fede8b4d7d31","Type":"ContainerStarted","Data":"c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.100881 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" event={"ID":"1b664cac-f1be-4b54-b85d-fede8b4d7d31","Type":"ContainerStarted","Data":"d7050c82a293de4981bb1136cb34984f9983b39afd30346c219c09e99e0451c4"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.125069 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.145712 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.146451 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.146501 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.146520 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.146547 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.146599 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.169194 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.190188 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.226662 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.249968 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.250105 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.250131 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.250169 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.250196 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.254155 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.276806 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.297824 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.316968 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.338033 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.353486 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.353557 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.353574 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.353601 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.353619 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.356321 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.377638 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.400554 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:29 crc kubenswrapper[4948]: E0220 08:06:29.400826 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.400763 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: E0220 08:06:29.400930 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs podName:62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd nodeName:}" failed. No retries permitted until 2026-02-20 08:06:30.400901223 +0000 UTC m=+39.375396083 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs") pod "network-metrics-daemon-xd86g" (UID: "62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.425835 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.444523 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.456687 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.456767 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.456792 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.456827 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.456854 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.462719 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.483847 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:29Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.559899 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.559960 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.560256 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.560291 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.560310 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.663185 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.663259 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.663283 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.663314 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.663342 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.676692 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 18:37:38.369965687 +0000 UTC Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.766639 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.766702 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.766722 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.766749 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.766771 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.870708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.870766 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.870788 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.870813 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.870832 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.974570 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.974635 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.974657 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.974684 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:29 crc kubenswrapper[4948]: I0220 08:06:29.974703 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:29Z","lastTransitionTime":"2026-02-20T08:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.078363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.078428 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.078445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.078470 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.078487 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.182199 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.182280 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.182293 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.182320 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.182335 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.285678 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.286031 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.286168 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.286310 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.286445 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.389518 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.389583 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.389604 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.389631 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.389652 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.413608 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:30 crc kubenswrapper[4948]: E0220 08:06:30.413784 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:30 crc kubenswrapper[4948]: E0220 08:06:30.414127 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs podName:62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd nodeName:}" failed. No retries permitted until 2026-02-20 08:06:32.414096922 +0000 UTC m=+41.388591782 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs") pod "network-metrics-daemon-xd86g" (UID: "62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.493052 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.493471 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.493605 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.493761 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.493894 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.598121 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.598192 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.598210 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.598238 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.598258 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.677707 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 21:13:07.062013621 +0000 UTC Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.701321 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.701391 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.701412 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.701439 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.701462 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.722425 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.722484 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:30 crc kubenswrapper[4948]: E0220 08:06:30.722647 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.722673 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.722715 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:30 crc kubenswrapper[4948]: E0220 08:06:30.722851 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:30 crc kubenswrapper[4948]: E0220 08:06:30.723013 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:30 crc kubenswrapper[4948]: E0220 08:06:30.723136 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.804752 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.804811 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.804830 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.804857 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.804876 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.908912 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.909017 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.909040 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.909072 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:30 crc kubenswrapper[4948]: I0220 08:06:30.909092 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:30Z","lastTransitionTime":"2026-02-20T08:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.013218 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.013313 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.013335 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.013361 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.013380 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.116200 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.116267 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.116285 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.116314 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.116335 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.220225 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.220308 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.220328 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.220364 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.220389 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.324128 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.324199 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.324219 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.324247 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.324269 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.428019 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.428079 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.428091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.428113 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.428130 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.531580 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.531681 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.531700 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.531728 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.531746 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.635190 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.635253 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.635271 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.635298 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.635318 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.678792 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 15:35:05.395643427 +0000 UTC Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.738320 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.738394 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.738413 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.738441 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.738460 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.741529 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.774070 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.796335 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.815093 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.840823 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.840877 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.840889 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.840910 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.840935 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.848040 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.873277 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.899259 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.920451 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.944128 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.944475 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.944541 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.944558 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.944584 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.944601 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:31Z","lastTransitionTime":"2026-02-20T08:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.963114 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.981468 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:31 crc kubenswrapper[4948]: I0220 08:06:31.998240 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:31Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.012953 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:32Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.041428 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:32Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.047451 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.047531 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.047551 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.047582 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.047602 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.064242 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:32Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.080689 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:32Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.100862 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:32Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.151648 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.151711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.151728 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.151753 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.151770 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.255726 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.255784 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.255803 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.255832 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.255850 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.359668 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.360087 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.360265 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.360321 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.360343 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.438658 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:32 crc kubenswrapper[4948]: E0220 08:06:32.438855 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:32 crc kubenswrapper[4948]: E0220 08:06:32.439000 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs podName:62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd nodeName:}" failed. No retries permitted until 2026-02-20 08:06:36.438943386 +0000 UTC m=+45.413438246 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs") pod "network-metrics-daemon-xd86g" (UID: "62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.463386 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.463469 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.463492 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.463525 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.463550 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.566692 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.566748 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.566765 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.566794 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.566812 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.670301 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.670372 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.670401 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.670438 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.670465 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.679253 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 17:06:39.819894113 +0000 UTC Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.721642 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.721669 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.721677 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.721841 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:32 crc kubenswrapper[4948]: E0220 08:06:32.722063 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:32 crc kubenswrapper[4948]: E0220 08:06:32.722237 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:32 crc kubenswrapper[4948]: E0220 08:06:32.722373 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:32 crc kubenswrapper[4948]: E0220 08:06:32.722523 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.732549 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.733782 4948 scope.go:117] "RemoveContainer" containerID="8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e" Feb 20 08:06:32 crc kubenswrapper[4948]: E0220 08:06:32.734072 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.773446 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.773509 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.773527 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.773551 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.773571 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.876934 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.877030 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.877052 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.877124 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.877144 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.980423 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.980485 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.980501 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.980525 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:32 crc kubenswrapper[4948]: I0220 08:06:32.980543 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:32Z","lastTransitionTime":"2026-02-20T08:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.084226 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.084313 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.084340 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.084371 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.084393 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.187824 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.187888 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.187912 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.187941 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.187963 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.291784 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.291835 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.291850 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.291873 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.291889 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.401373 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.401436 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.401454 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.401484 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.401502 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.505029 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.505106 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.505123 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.505151 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.505170 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.608886 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.608954 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.609009 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.609039 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.609061 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.680127 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 13:44:21.15647115 +0000 UTC Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.711883 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.711961 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.712012 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.712041 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.712060 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.815706 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.815798 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.815818 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.815844 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.815864 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.919213 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.919274 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.919292 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.919320 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:33 crc kubenswrapper[4948]: I0220 08:06:33.919348 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:33Z","lastTransitionTime":"2026-02-20T08:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.022318 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.022378 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.022396 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.022423 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.022445 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.124767 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.124881 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.124907 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.124937 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.124958 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.229231 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.229318 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.229347 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.229381 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.229406 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.332732 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.332777 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.332794 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.332814 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.332832 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.436424 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.436513 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.436529 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.436553 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.436569 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.540231 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.540293 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.540310 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.540337 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.540357 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.643764 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.643844 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.643867 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.643895 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.643919 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.681136 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 15:02:47.403721519 +0000 UTC Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.721478 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.721543 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.721499 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.721479 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:34 crc kubenswrapper[4948]: E0220 08:06:34.721744 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:34 crc kubenswrapper[4948]: E0220 08:06:34.721895 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:34 crc kubenswrapper[4948]: E0220 08:06:34.722106 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:34 crc kubenswrapper[4948]: E0220 08:06:34.722308 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.746847 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.746917 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.746943 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.747011 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.747062 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.850541 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.850597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.850610 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.850631 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.850644 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.953265 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.953329 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.953347 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.953374 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:34 crc kubenswrapper[4948]: I0220 08:06:34.953393 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:34Z","lastTransitionTime":"2026-02-20T08:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.056809 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.056884 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.056906 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.056938 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.056958 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.159516 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.159641 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.159666 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.159704 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.159723 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.263243 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.263317 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.263336 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.263363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.263384 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.365726 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.365773 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.365785 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.365803 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.365813 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.468931 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.469077 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.469092 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.469112 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.469124 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.571944 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.571997 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.572008 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.572026 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.572038 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.675297 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.675344 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.675356 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.675381 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.675400 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.681984 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 03:58:01.055675842 +0000 UTC Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.778040 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.778107 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.778132 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.778159 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.778179 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.881865 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.881924 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.881937 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.881960 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.882001 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.984640 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.984864 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.984886 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.984913 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:35 crc kubenswrapper[4948]: I0220 08:06:35.984931 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:35Z","lastTransitionTime":"2026-02-20T08:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.088019 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.088089 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.088106 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.088134 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.088155 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.191886 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.191953 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.192012 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.192047 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.192071 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.295220 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.295319 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.295340 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.295363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.295383 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.398275 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.398335 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.398352 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.398381 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.398400 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.495565 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:36 crc kubenswrapper[4948]: E0220 08:06:36.495754 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:36 crc kubenswrapper[4948]: E0220 08:06:36.495812 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs podName:62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd nodeName:}" failed. No retries permitted until 2026-02-20 08:06:44.495796244 +0000 UTC m=+53.470291064 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs") pod "network-metrics-daemon-xd86g" (UID: "62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.501594 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.501631 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.501643 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.501662 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.501676 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.604998 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.605066 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.605090 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.605118 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.605141 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.683121 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 08:58:33.057202881 +0000 UTC Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.708054 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.708117 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.708138 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.708163 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.708181 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.721597 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.721609 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:36 crc kubenswrapper[4948]: E0220 08:06:36.721755 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.721632 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.721631 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:36 crc kubenswrapper[4948]: E0220 08:06:36.721926 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:36 crc kubenswrapper[4948]: E0220 08:06:36.722120 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:36 crc kubenswrapper[4948]: E0220 08:06:36.722267 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.811000 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.811062 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.811078 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.811102 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.811120 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.914138 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.914204 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.914230 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.914263 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:36 crc kubenswrapper[4948]: I0220 08:06:36.914289 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:36Z","lastTransitionTime":"2026-02-20T08:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.016942 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.017098 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.017121 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.017151 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.017188 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.119891 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.119960 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.120016 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.120051 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.120074 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.222922 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.223007 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.223028 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.223053 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.223072 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.326833 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.326897 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.326914 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.326940 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.326958 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.429727 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.429767 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.429778 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.429795 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.429807 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.532648 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.532718 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.532737 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.532763 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.532781 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.636587 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.636636 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.636645 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.636660 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.636671 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.683312 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 03:19:44.299217979 +0000 UTC Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.739861 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.739937 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.739962 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.740049 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.740078 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.843179 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.843235 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.843253 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.843276 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.843292 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.945920 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.946018 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.946030 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.946056 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:37 crc kubenswrapper[4948]: I0220 08:06:37.946072 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:37Z","lastTransitionTime":"2026-02-20T08:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.049043 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.049128 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.049143 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.049170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.049186 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.151458 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.151830 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.152057 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.152237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.152391 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.255696 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.255747 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.255764 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.255783 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.255794 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.359497 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.359557 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.359570 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.359592 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.359607 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.462697 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.462750 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.462765 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.462785 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.462799 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.566144 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.566229 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.566267 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.566304 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.566330 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.669578 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.669669 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.669693 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.669728 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.669752 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.683842 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 21:33:02.940916729 +0000 UTC Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.722252 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.722369 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.722431 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:38 crc kubenswrapper[4948]: E0220 08:06:38.722479 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.722390 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:38 crc kubenswrapper[4948]: E0220 08:06:38.722582 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:38 crc kubenswrapper[4948]: E0220 08:06:38.722800 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:38 crc kubenswrapper[4948]: E0220 08:06:38.722932 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.772962 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.773071 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.773102 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.773132 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.773154 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.876800 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.876843 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.876854 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.876871 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.876885 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.980010 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.980097 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.980160 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.980194 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:38 crc kubenswrapper[4948]: I0220 08:06:38.980217 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:38Z","lastTransitionTime":"2026-02-20T08:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.057735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.057791 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.057803 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.057823 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.057837 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: E0220 08:06:39.080043 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:39Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.085533 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.085593 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.085626 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.085653 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.085671 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: E0220 08:06:39.106520 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:39Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.111501 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.111558 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.111576 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.111600 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.111621 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: E0220 08:06:39.131269 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:39Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.137022 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.137175 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.137288 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.137340 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.137403 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: E0220 08:06:39.160211 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:39Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.166825 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.166887 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.166907 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.166933 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.166952 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: E0220 08:06:39.188484 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:39Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:39 crc kubenswrapper[4948]: E0220 08:06:39.188719 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.191094 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.191149 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.191161 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.191182 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.191198 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.295116 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.295701 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.296032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.296242 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.296449 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.400714 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.400773 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.400794 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.400825 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.400849 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.504030 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.504092 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.504116 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.504148 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.504175 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.606923 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.607015 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.607033 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.607060 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.607079 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.684913 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 11:36:56.624396106 +0000 UTC Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.709430 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.709500 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.709528 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.709558 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.709578 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.813324 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.813408 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.813427 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.813454 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.813475 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.916956 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.917060 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.917084 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.917118 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:39 crc kubenswrapper[4948]: I0220 08:06:39.917143 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:39Z","lastTransitionTime":"2026-02-20T08:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.020524 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.020591 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.020610 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.020638 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.020658 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.124461 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.124533 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.124551 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.124577 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.124598 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.233666 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.233751 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.233775 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.233810 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.233835 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.336597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.336650 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.336661 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.336680 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.336693 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.439026 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.439094 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.439117 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.439146 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.439168 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.543623 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.543707 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.543732 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.543767 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.543790 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.648099 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.648176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.648200 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.648231 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.648253 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.685625 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 03:29:19.944070432 +0000 UTC Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.722357 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.722399 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.722554 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:40 crc kubenswrapper[4948]: E0220 08:06:40.722546 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.722698 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:40 crc kubenswrapper[4948]: E0220 08:06:40.722836 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:40 crc kubenswrapper[4948]: E0220 08:06:40.723071 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:40 crc kubenswrapper[4948]: E0220 08:06:40.723203 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.751628 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.752053 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.752245 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.752439 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.752636 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.856797 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.856878 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.856901 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.856929 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.856950 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.960562 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.960661 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.960685 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.960719 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:40 crc kubenswrapper[4948]: I0220 08:06:40.960743 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:40Z","lastTransitionTime":"2026-02-20T08:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.064301 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.064367 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.064384 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.064409 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.064437 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.168094 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.168183 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.168202 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.168226 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.168246 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.271488 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.271783 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.271933 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.272138 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.272283 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.375474 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.375542 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.375561 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.375590 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.375615 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.478616 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.478691 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.478709 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.478736 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.478756 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.581862 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.581951 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.582005 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.582032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.582051 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.685253 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.685331 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.685350 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.685382 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.685403 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.686076 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 23:42:08.028612677 +0000 UTC Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.745814 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.763931 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.781706 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.787861 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.787935 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.787954 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.788013 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.788035 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.798465 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.840191 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.866581 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.884541 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.890796 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.890871 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.890894 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.890948 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.891003 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.911467 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.932666 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:41Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.994637 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.994764 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.994787 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.994823 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:41 crc kubenswrapper[4948]: I0220 08:06:41.994855 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:41Z","lastTransitionTime":"2026-02-20T08:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.012248 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:42Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.042787 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:42Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.055021 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:42Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.076321 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:42Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.088102 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:42Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.098418 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.098470 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.098482 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.098502 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.098516 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.102636 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:42Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.114881 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:42Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.128603 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:42Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.202346 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.202415 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.202432 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.202457 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.202476 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.305331 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.305393 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.305405 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.305425 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.305745 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.408648 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.408710 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.408728 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.408757 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.408778 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.512282 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.512364 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.512387 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.512414 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.512441 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.615510 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.615596 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.615633 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.615667 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.615698 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.687105 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 21:55:29.672665013 +0000 UTC Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.718900 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.718967 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.719022 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.719051 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.719071 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.721727 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.721822 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.721842 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.721727 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:42 crc kubenswrapper[4948]: E0220 08:06:42.721936 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:42 crc kubenswrapper[4948]: E0220 08:06:42.722118 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:42 crc kubenswrapper[4948]: E0220 08:06:42.722291 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:42 crc kubenswrapper[4948]: E0220 08:06:42.722496 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.822435 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.822511 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.822530 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.822557 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.822578 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.925467 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.925557 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.925591 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.925620 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:42 crc kubenswrapper[4948]: I0220 08:06:42.925638 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:42Z","lastTransitionTime":"2026-02-20T08:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.028770 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.028827 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.028880 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.028911 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.029519 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.153907 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.153967 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.154027 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.154064 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.154087 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.258201 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.258286 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.258307 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.258344 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.258367 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.362384 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.362484 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.362503 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.362529 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.362549 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.465661 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.465731 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.465751 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.465783 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.465811 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.568541 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.568651 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.568671 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.568699 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.568718 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.671227 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.671374 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.671399 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.671422 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.671443 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.688157 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 16:28:12.724100466 +0000 UTC Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.723382 4948 scope.go:117] "RemoveContainer" containerID="8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.779536 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.779599 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.779625 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.779656 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.779676 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.883083 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.884155 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.884210 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.884248 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:43 crc kubenswrapper[4948]: I0220 08:06:43.884274 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:43.991472 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:43.991524 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:43.991542 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:43.991565 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:43.991583 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:43Z","lastTransitionTime":"2026-02-20T08:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.095909 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.095962 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.096021 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.096050 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.096071 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.161848 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/1.log" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.166817 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.167535 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.196235 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.198756 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.198809 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.198835 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.198869 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.198894 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.216466 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.241752 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.265492 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.292837 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.302162 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.302213 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.302230 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.302249 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.302261 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.313284 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.339388 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.364078 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.405473 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.405572 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.405591 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.405613 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.405626 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.406306 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.433721 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.454460 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.474657 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.491616 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.491937 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.491790 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.491590 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.492105 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.492347 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:07:16.492324771 +0000 UTC m=+85.466819601 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.492496 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:07:16.492432243 +0000 UTC m=+85.466927073 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.508517 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.508580 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.508595 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.508618 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.508636 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.518450 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.533774 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.550204 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.568902 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.593367 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.593653 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:07:16.593604247 +0000 UTC m=+85.568099107 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.593766 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.593911 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.593951 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.593999 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.594016 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.594041 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.594078 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 08:07:16.594057087 +0000 UTC m=+85.568551917 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.594163 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.594187 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.594315 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.594335 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.594319 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs podName:62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd nodeName:}" failed. No retries permitted until 2026-02-20 08:07:00.594274323 +0000 UTC m=+69.568769333 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs") pod "network-metrics-daemon-xd86g" (UID: "62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.594395 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 08:07:16.594384945 +0000 UTC m=+85.568879765 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.611406 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.611442 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.611453 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.611484 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.611503 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.688924 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 16:49:01.154887844 +0000 UTC Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.714445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.714491 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.714500 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.714518 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.714530 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.721794 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.721830 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.721863 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.721909 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.722049 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.721794 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.722120 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:44 crc kubenswrapper[4948]: E0220 08:06:44.722169 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.818120 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.818183 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.818199 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.818228 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.818248 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.920917 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.921076 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.921105 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.921142 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:44 crc kubenswrapper[4948]: I0220 08:06:44.921173 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:44Z","lastTransitionTime":"2026-02-20T08:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.025379 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.025437 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.025449 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.025469 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.025481 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.129213 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.129570 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.129700 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.129894 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.130056 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.175530 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/2.log" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.177813 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/1.log" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.186209 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf" exitCode=1 Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.186336 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.186692 4948 scope.go:117] "RemoveContainer" containerID="8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.187693 4948 scope.go:117] "RemoveContainer" containerID="bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf" Feb 20 08:06:45 crc kubenswrapper[4948]: E0220 08:06:45.188035 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.208740 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.229451 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.234551 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.234613 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.234630 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.234657 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.234682 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.274057 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.294893 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.316378 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.334564 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.338552 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.338629 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.338657 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.338681 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.338703 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.346455 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.350266 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.367760 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.400509 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.423180 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.442233 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.442355 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.442384 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.442416 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.442447 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.444015 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.465008 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.490351 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.513940 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.536153 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.545457 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.545544 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.545568 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.545599 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.545623 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.559546 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.581310 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.600516 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.621056 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.642052 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.648465 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.648511 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.648522 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.648538 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.648551 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.662755 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.685050 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.689050 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 23:52:12.695636297 +0000 UTC Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.722136 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.742348 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.752037 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.752104 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.752126 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.752152 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.752177 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.761769 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.778388 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.793263 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.826916 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8aef60f26bd969af4082460f139139643b989d57f366e6d9c2dd743e8d9b1f2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:26Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:141\\\\nI0220 08:06:26.064417 6421 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:06:26.065536 6421 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 08:06:26.065606 6421 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 08:06:26.065641 6421 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 08:06:26.065650 6421 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 08:06:26.065682 6421 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:06:26.065678 6421 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 08:06:26.065713 6421 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 08:06:26.065690 6421 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:06:26.065709 6421 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 08:06:26.065775 6421 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 08:06:26.065788 6421 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 08:06:26.065794 6421 factory.go:656] Stopping watch factory\\\\nI0220 08:06:26.065810 6421 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.850420 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.855442 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.855542 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.855569 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.855605 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.855628 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.876491 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.897314 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.916821 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.937667 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.953887 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.961345 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.961382 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.961394 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.961414 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.961428 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:45Z","lastTransitionTime":"2026-02-20T08:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.975731 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:45 crc kubenswrapper[4948]: I0220 08:06:45.991512 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:45Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.065517 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.065578 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.065590 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.065613 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.065627 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.170209 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.170283 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.170302 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.170332 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.170351 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.194212 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/2.log" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.200771 4948 scope.go:117] "RemoveContainer" containerID="bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf" Feb 20 08:06:46 crc kubenswrapper[4948]: E0220 08:06:46.201108 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.218474 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.236593 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.274039 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.274091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.274104 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.274129 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.274145 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.278554 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.301196 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.318736 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.344385 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.362795 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.377035 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.377099 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.377120 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.377148 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.377170 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.396861 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.419941 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.436070 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.452946 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.476276 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.480387 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.480433 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.480449 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.480473 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.480489 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.496561 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.517474 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.536131 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.554702 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.575267 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.583471 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.583595 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.583611 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.583633 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.583646 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.593394 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:46Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.686808 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.686891 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.686911 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.686950 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.687011 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.689967 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 18:24:17.177828311 +0000 UTC Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.721434 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.721494 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.721454 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.721434 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:46 crc kubenswrapper[4948]: E0220 08:06:46.721602 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:46 crc kubenswrapper[4948]: E0220 08:06:46.721828 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:46 crc kubenswrapper[4948]: E0220 08:06:46.721966 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:46 crc kubenswrapper[4948]: E0220 08:06:46.722111 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.790531 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.790602 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.790621 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.790646 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.790661 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.893853 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.893924 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.893944 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.894003 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.894022 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.996795 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.996851 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.996867 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.996893 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:46 crc kubenswrapper[4948]: I0220 08:06:46.996910 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:46Z","lastTransitionTime":"2026-02-20T08:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.100662 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.100727 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.100751 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.100794 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.100815 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.203821 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.203891 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.203915 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.203945 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.204016 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.307305 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.307431 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.307452 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.307476 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.307495 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.410736 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.410804 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.410827 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.410858 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.410881 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.514152 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.514237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.514258 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.514285 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.514303 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.617674 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.617745 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.617763 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.617790 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.617809 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.691104 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 06:26:23.148617268 +0000 UTC Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.720831 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.720882 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.720901 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.720929 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.720949 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.829579 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.829662 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.829680 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.829711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.829729 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.933655 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.933726 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.933746 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.933777 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:47 crc kubenswrapper[4948]: I0220 08:06:47.933797 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:47Z","lastTransitionTime":"2026-02-20T08:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.037291 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.037392 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.037412 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.037445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.037465 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.141304 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.141365 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.141383 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.141409 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.141431 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.245191 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.245236 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.245250 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.245269 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.245284 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.349334 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.349436 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.349454 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.349535 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.349557 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.452864 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.452955 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.453028 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.453064 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.453091 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.556947 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.557137 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.557170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.557203 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.557227 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.660780 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.660844 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.660862 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.660887 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.660906 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.691887 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 01:30:44.263786704 +0000 UTC Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.722655 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.722704 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.722665 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.722656 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:48 crc kubenswrapper[4948]: E0220 08:06:48.722812 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:48 crc kubenswrapper[4948]: E0220 08:06:48.723024 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:48 crc kubenswrapper[4948]: E0220 08:06:48.723158 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:48 crc kubenswrapper[4948]: E0220 08:06:48.723304 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.764649 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.764741 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.764767 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.764800 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.764825 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.868296 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.868382 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.868401 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.868459 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.868480 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.972283 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.972727 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.972827 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.972939 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:48 crc kubenswrapper[4948]: I0220 08:06:48.973062 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:48Z","lastTransitionTime":"2026-02-20T08:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.076117 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.076158 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.076170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.076187 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.076202 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.180332 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.180411 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.180433 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.180467 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.180489 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.283496 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.283546 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.283566 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.283588 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.283610 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.386282 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.386343 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.386367 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.386394 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.386417 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.399619 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.399777 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.399805 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.399832 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.399853 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: E0220 08:06:49.421716 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:49Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.428182 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.428245 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.428268 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.428295 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.428316 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: E0220 08:06:49.451192 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:49Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.457599 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.457654 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.457671 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.457699 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.457718 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: E0220 08:06:49.479788 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:49Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.485126 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.485170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.485197 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.485229 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.485255 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: E0220 08:06:49.504939 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:49Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.510014 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.510055 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.510073 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.510096 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.510114 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: E0220 08:06:49.528740 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:49Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:49 crc kubenswrapper[4948]: E0220 08:06:49.529127 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.530916 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.531025 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.531044 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.531064 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.531081 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.634355 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.634426 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.634446 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.634474 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.634495 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.693065 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 13:48:21.287241825 +0000 UTC Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.737319 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.737404 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.737431 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.737462 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.737487 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.840577 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.840642 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.840661 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.840689 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.840711 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.944186 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.944254 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.944278 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.944308 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:49 crc kubenswrapper[4948]: I0220 08:06:49.944331 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:49Z","lastTransitionTime":"2026-02-20T08:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.047520 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.047580 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.047603 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.047629 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.047649 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.150832 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.150900 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.150923 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.150949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.151007 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.253319 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.253392 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.253411 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.253440 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.253460 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.356637 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.356702 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.356731 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.356763 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.356786 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.460260 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.460326 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.460339 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.460358 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.460370 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.563494 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.563552 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.563570 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.563593 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.563611 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.666252 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.666457 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.666483 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.666506 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.666523 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.694221 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 02:06:40.87345129 +0000 UTC Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.721837 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.721906 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.721871 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.721922 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:50 crc kubenswrapper[4948]: E0220 08:06:50.722088 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:50 crc kubenswrapper[4948]: E0220 08:06:50.722409 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:50 crc kubenswrapper[4948]: E0220 08:06:50.722524 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:50 crc kubenswrapper[4948]: E0220 08:06:50.722658 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.769520 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.769611 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.769638 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.769672 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.769699 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.873562 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.873619 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.873635 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.873659 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.873675 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.977040 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.977118 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.977136 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.977167 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:50 crc kubenswrapper[4948]: I0220 08:06:50.977191 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:50Z","lastTransitionTime":"2026-02-20T08:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.080480 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.080816 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.081072 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.081226 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.081377 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.184177 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.184258 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.184276 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.184301 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.184319 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.287583 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.287678 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.287708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.287740 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.287759 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.391591 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.391654 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.391675 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.391704 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.391726 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.495622 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.495695 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.495715 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.495747 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.495764 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.599451 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.599673 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.599697 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.599723 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.599741 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.695130 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 01:55:40.955891129 +0000 UTC Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.703099 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.703319 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.703476 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.703630 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.703771 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.760765 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.782739 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.799904 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.806729 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.806803 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.806822 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.806853 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.806872 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.818374 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.838055 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.862754 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.882920 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.908803 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.911836 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.911902 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.911922 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.911949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.911968 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:51Z","lastTransitionTime":"2026-02-20T08:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.929101 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.966762 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:51 crc kubenswrapper[4948]: I0220 08:06:51.984200 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:51Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.003135 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:52Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.016654 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.016721 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.016752 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.016786 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.016805 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.027475 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:52Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.045314 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:52Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.063306 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:52Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.081737 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:52Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.098555 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:52Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.115289 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:52Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.119588 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.119662 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.119683 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.119711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.119730 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.223348 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.223443 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.223464 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.223498 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.223519 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.327013 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.327068 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.327086 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.327112 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.327175 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.430078 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.430144 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.430162 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.430187 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.430210 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.534116 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.534180 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.534245 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.534273 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.534295 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.637744 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.637792 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.637810 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.637837 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.637855 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.696249 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 08:47:54.079079862 +0000 UTC Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.722124 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.722167 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.722299 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.722395 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:52 crc kubenswrapper[4948]: E0220 08:06:52.722389 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:52 crc kubenswrapper[4948]: E0220 08:06:52.722584 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:52 crc kubenswrapper[4948]: E0220 08:06:52.722790 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:52 crc kubenswrapper[4948]: E0220 08:06:52.723029 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.741295 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.741357 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.741375 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.741401 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.741420 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.844849 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.844925 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.844943 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.845006 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.845026 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.948502 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.949026 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.949219 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.949383 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:52 crc kubenswrapper[4948]: I0220 08:06:52.949516 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:52Z","lastTransitionTime":"2026-02-20T08:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.053556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.053602 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.053617 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.053638 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.053653 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.157204 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.157276 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.157295 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.157323 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.157342 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.261047 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.261527 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.261734 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.261926 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.262159 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.366388 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.366458 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.366478 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.366507 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.366526 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.470016 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.470328 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.470508 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.470624 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.470713 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.573911 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.574073 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.574107 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.574144 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.574171 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.677901 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.678025 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.678046 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.678081 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.678102 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.697475 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 18:22:32.438410222 +0000 UTC Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.782317 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.782404 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.782445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.782480 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.782502 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.885381 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.885437 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.885453 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.885475 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.885492 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.988623 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.988682 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.988698 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.988723 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:53 crc kubenswrapper[4948]: I0220 08:06:53.988741 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:53Z","lastTransitionTime":"2026-02-20T08:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.092031 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.092090 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.092117 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.092139 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.092155 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.194766 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.194855 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.194878 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.194906 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.194923 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.298109 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.298274 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.298299 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.298326 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.298344 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.402949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.403041 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.403061 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.403088 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.403109 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.506815 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.506868 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.506885 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.506913 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.506933 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.610967 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.611068 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.611090 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.611117 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.611138 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.697994 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 01:45:21.545583545 +0000 UTC Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.714927 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.715047 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.715074 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.715106 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.715130 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.722202 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.722259 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.722290 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.722202 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:54 crc kubenswrapper[4948]: E0220 08:06:54.722402 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:54 crc kubenswrapper[4948]: E0220 08:06:54.722500 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:54 crc kubenswrapper[4948]: E0220 08:06:54.722625 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:54 crc kubenswrapper[4948]: E0220 08:06:54.722746 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.818708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.818775 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.818798 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.818829 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.818851 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.922459 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.922515 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.922532 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.922556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:54 crc kubenswrapper[4948]: I0220 08:06:54.922577 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:54Z","lastTransitionTime":"2026-02-20T08:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.025585 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.025662 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.025680 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.025708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.025727 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.128928 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.129034 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.129057 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.129082 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.129104 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.232527 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.232595 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.232615 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.232640 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.232659 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.336493 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.336556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.336577 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.336601 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.336621 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.438906 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.438952 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.438963 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.438997 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.439010 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.541652 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.541761 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.541785 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.541810 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.541828 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.645383 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.645430 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.645446 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.645464 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.645476 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.698316 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 00:11:58.672805079 +0000 UTC Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.748449 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.748522 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.748545 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.748576 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.748595 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.850999 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.851051 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.851063 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.851082 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.851095 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.954071 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.954131 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.954141 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.954159 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:55 crc kubenswrapper[4948]: I0220 08:06:55.954169 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:55Z","lastTransitionTime":"2026-02-20T08:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.056749 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.056820 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.056839 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.056869 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.056891 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.161480 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.161542 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.161555 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.161578 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.161595 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.265014 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.265065 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.265075 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.265407 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.265422 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.369563 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.369665 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.369684 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.369709 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.369727 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.473277 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.473690 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.473865 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.474091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.474308 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.577540 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.577634 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.577651 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.577674 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.577693 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.680963 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.681031 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.681041 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.681057 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.681070 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.699393 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 23:16:06.111703225 +0000 UTC Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.721953 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:56 crc kubenswrapper[4948]: E0220 08:06:56.722167 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.722256 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.722346 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:56 crc kubenswrapper[4948]: E0220 08:06:56.722373 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.722460 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:56 crc kubenswrapper[4948]: E0220 08:06:56.722558 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:56 crc kubenswrapper[4948]: E0220 08:06:56.722696 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.784127 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.784162 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.784170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.784184 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.784194 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.886653 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.886711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.886729 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.886773 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.886792 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.989170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.989241 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.989264 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.989293 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:56 crc kubenswrapper[4948]: I0220 08:06:56.989326 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:56Z","lastTransitionTime":"2026-02-20T08:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.092919 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.093008 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.093018 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.093040 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.093051 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.195928 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.196009 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.196026 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.196048 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.196065 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.299103 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.299156 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.299166 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.299186 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.299197 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.402711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.402776 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.402820 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.402846 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.402863 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.506423 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.506478 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.506497 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.506521 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.506539 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.610299 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.610344 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.610360 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.610384 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.610400 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.699615 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 12:05:56.474940802 +0000 UTC Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.713723 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.713817 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.713845 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.713880 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.713906 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.817278 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.817334 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.817344 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.817359 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.817369 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.920639 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.920707 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.920732 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.920757 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:57 crc kubenswrapper[4948]: I0220 08:06:57.920776 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:57Z","lastTransitionTime":"2026-02-20T08:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.023489 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.023554 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.023578 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.023607 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.023631 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.126701 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.126763 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.126786 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.126812 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.126831 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.230685 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.230744 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.230762 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.230787 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.230806 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.360712 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.360749 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.360760 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.360779 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.360793 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.464857 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.464920 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.464934 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.464952 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.464989 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.568205 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.568246 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.568255 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.568274 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.568284 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.671670 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.671717 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.671727 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.671745 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.671755 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.699920 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 00:49:32.57491579 +0000 UTC Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.722338 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.722340 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.722368 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:06:58 crc kubenswrapper[4948]: E0220 08:06:58.722539 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.722553 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:06:58 crc kubenswrapper[4948]: E0220 08:06:58.722653 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:06:58 crc kubenswrapper[4948]: E0220 08:06:58.722734 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:06:58 crc kubenswrapper[4948]: E0220 08:06:58.722796 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.734676 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.773636 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.773672 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.773680 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.773693 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.773703 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.886016 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.886100 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.886127 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.886159 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.886183 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.988481 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.988560 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.988576 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.988596 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:58 crc kubenswrapper[4948]: I0220 08:06:58.988610 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:58Z","lastTransitionTime":"2026-02-20T08:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.091017 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.091093 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.091105 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.091126 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.091142 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.193220 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.193251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.193263 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.193280 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.193290 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.295094 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.295135 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.295147 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.295162 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.295186 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.397927 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.398030 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.398046 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.398065 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.398080 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.501057 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.501099 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.501108 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.501125 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.501135 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.592211 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.592255 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.592264 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.592280 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.592291 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: E0220 08:06:59.611307 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:59Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.615318 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.615374 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.615389 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.615411 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.615428 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: E0220 08:06:59.626858 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:59Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.631680 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.631730 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.631740 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.631755 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.631765 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: E0220 08:06:59.644775 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:59Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.650119 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.650235 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.650254 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.650280 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.650301 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: E0220 08:06:59.669634 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:59Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.674678 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.674758 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.674772 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.674794 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.674874 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: E0220 08:06:59.688111 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:59Z is after 2025-08-24T17:21:41Z" Feb 20 08:06:59 crc kubenswrapper[4948]: E0220 08:06:59.688239 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.689856 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.689898 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.689912 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.689931 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.689945 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.700288 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 13:19:15.391288625 +0000 UTC Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.792590 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.792663 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.792679 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.792700 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.792719 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.894738 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.894777 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.894808 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.894822 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.894832 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.997226 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.997269 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.997283 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.997297 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:06:59 crc kubenswrapper[4948]: I0220 08:06:59.997311 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:06:59Z","lastTransitionTime":"2026-02-20T08:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.100253 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.100301 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.100314 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.100337 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.100352 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.203225 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.203294 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.203318 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.203349 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.203372 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.305935 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.305992 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.306005 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.306026 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.306040 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.408894 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.408927 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.408939 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.408954 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.408966 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.511735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.511766 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.511777 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.511790 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.511799 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.614310 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.614370 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.614391 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.614422 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.614442 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.681295 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:00 crc kubenswrapper[4948]: E0220 08:07:00.681498 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:07:00 crc kubenswrapper[4948]: E0220 08:07:00.681588 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs podName:62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd nodeName:}" failed. No retries permitted until 2026-02-20 08:07:32.681563292 +0000 UTC m=+101.656058112 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs") pod "network-metrics-daemon-xd86g" (UID: "62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.700597 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 23:56:33.573767755 +0000 UTC Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.716769 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.716807 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.716818 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.716835 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.716848 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.722386 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.722449 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:00 crc kubenswrapper[4948]: E0220 08:07:00.722529 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.722629 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:00 crc kubenswrapper[4948]: E0220 08:07:00.722776 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.722850 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:00 crc kubenswrapper[4948]: E0220 08:07:00.722949 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:00 crc kubenswrapper[4948]: E0220 08:07:00.723477 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.723935 4948 scope.go:117] "RemoveContainer" containerID="bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf" Feb 20 08:07:00 crc kubenswrapper[4948]: E0220 08:07:00.724195 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.821599 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.821665 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.821686 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.821714 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.821752 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.924158 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.924208 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.924222 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.924244 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:00 crc kubenswrapper[4948]: I0220 08:07:00.924257 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:00Z","lastTransitionTime":"2026-02-20T08:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.027735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.027782 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.027796 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.027817 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.027828 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.131245 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.131292 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.131301 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.131317 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.131327 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.233597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.233631 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.233642 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.233656 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.233666 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.335944 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.336005 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.336014 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.336032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.336044 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.438488 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.438522 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.438535 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.438556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.438568 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.540839 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.540878 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.540892 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.540911 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.540926 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.643623 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.643656 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.643666 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.643681 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.643691 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.701410 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 17:25:14.277958375 +0000 UTC Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.741926 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.745501 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.745521 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.745529 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.745539 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.745547 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.761702 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.783660 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.801026 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.820763 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.839963 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.848340 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.848396 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.848409 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.848433 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.848450 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.859063 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.881729 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.915128 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.934591 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.949645 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.950758 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.950809 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.950820 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.950837 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.950847 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:01Z","lastTransitionTime":"2026-02-20T08:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.963839 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.974662 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.983906 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ad7145-074b-49df-a0e9-cfec6e3974ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:01 crc kubenswrapper[4948]: I0220 08:07:01.998507 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:01Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.009269 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.025953 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.040135 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.052938 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.053001 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.053015 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.053031 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.053045 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.067953 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.154944 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.155018 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.155035 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.155056 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.155074 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.258986 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.259358 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.259367 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.259385 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.259397 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.262493 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/0.log" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.262539 4948 generic.go:334] "Generic (PLEG): container finished" podID="b96124e4-0a74-4578-9142-fd728eb9f99e" containerID="34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e" exitCode=1 Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.262583 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7s56" event={"ID":"b96124e4-0a74-4578-9142-fd728eb9f99e","Type":"ContainerDied","Data":"34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.263583 4948 scope.go:117] "RemoveContainer" containerID="34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.287623 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.305246 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.319213 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.331768 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.345649 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.359243 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.368243 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.368309 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.368328 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.368357 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.368378 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.372728 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.392939 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"2026-02-20T08:06:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45\\\\n2026-02-20T08:06:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45 to /host/opt/cni/bin/\\\\n2026-02-20T08:06:17Z [verbose] multus-daemon started\\\\n2026-02-20T08:06:17Z [verbose] Readiness Indicator file check\\\\n2026-02-20T08:07:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.413898 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.427925 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.441785 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.457595 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.469396 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.471298 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.471337 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.471345 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.471363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.471377 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.480287 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ad7145-074b-49df-a0e9-cfec6e3974ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.494270 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.504869 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.517265 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.533515 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.558846 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:02Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.573850 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.573905 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.573924 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.573950 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.573996 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.677653 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.677712 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.677729 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.677749 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.677761 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.702288 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 23:09:13.973968472 +0000 UTC Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.721890 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.721930 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.721904 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.721894 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:02 crc kubenswrapper[4948]: E0220 08:07:02.722076 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:02 crc kubenswrapper[4948]: E0220 08:07:02.722220 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:02 crc kubenswrapper[4948]: E0220 08:07:02.722390 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:02 crc kubenswrapper[4948]: E0220 08:07:02.722523 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.780360 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.780424 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.780443 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.780471 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.780490 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.883899 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.883949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.883958 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.883988 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.884000 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.986915 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.987265 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.987328 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.987393 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:02 crc kubenswrapper[4948]: I0220 08:07:02.987460 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:02Z","lastTransitionTime":"2026-02-20T08:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.090710 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.090805 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.090831 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.090865 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.090893 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.194921 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.195106 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.195225 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.196131 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.196196 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.269939 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/0.log" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.270042 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7s56" event={"ID":"b96124e4-0a74-4578-9142-fd728eb9f99e","Type":"ContainerStarted","Data":"2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.288715 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.298754 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.298807 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.298831 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.298858 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.298878 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.315243 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.331629 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.344559 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.358511 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.374803 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.394186 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"2026-02-20T08:06:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45\\\\n2026-02-20T08:06:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45 to /host/opt/cni/bin/\\\\n2026-02-20T08:06:17Z [verbose] multus-daemon started\\\\n2026-02-20T08:06:17Z [verbose] Readiness Indicator file check\\\\n2026-02-20T08:07:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.401156 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.401189 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.401220 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.401237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.401250 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.413661 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.430006 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.441241 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.454517 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.464803 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.487129 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.501749 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.503371 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.503424 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.503444 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.503467 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.503484 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.515344 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.533474 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.545487 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.567827 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.581449 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ad7145-074b-49df-a0e9-cfec6e3974ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:03Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.606083 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.606174 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.606192 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.606219 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.606237 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.703083 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 18:34:25.93003713 +0000 UTC Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.709036 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.709074 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.709084 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.709101 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.709114 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.811301 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.811365 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.811379 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.811398 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.811411 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.915017 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.915067 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.915076 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.915091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:03 crc kubenswrapper[4948]: I0220 08:07:03.915104 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:03Z","lastTransitionTime":"2026-02-20T08:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.018568 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.018637 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.018657 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.018680 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.018695 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:04Z","lastTransitionTime":"2026-02-20T08:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.121888 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.121947 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.121962 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.122000 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.122012 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:04Z","lastTransitionTime":"2026-02-20T08:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.224878 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.224923 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.224936 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.224955 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.224988 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:04Z","lastTransitionTime":"2026-02-20T08:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.326881 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.326923 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.326933 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.326950 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.326963 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:04Z","lastTransitionTime":"2026-02-20T08:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.429949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.430584 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.430687 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.430845 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.430957 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:04Z","lastTransitionTime":"2026-02-20T08:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.533934 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.534017 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.534036 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.534061 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.534080 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:04Z","lastTransitionTime":"2026-02-20T08:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.994133 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.994171 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:04 crc kubenswrapper[4948]: E0220 08:07:04.994271 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.994133 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:04 crc kubenswrapper[4948]: E0220 08:07:04.994359 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.994425 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:04 crc kubenswrapper[4948]: E0220 08:07:04.994539 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:04 crc kubenswrapper[4948]: E0220 08:07:04.994599 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.994639 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 13:09:44.211385592 +0000 UTC Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.995343 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.995362 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.995372 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.995386 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:04 crc kubenswrapper[4948]: I0220 08:07:04.995397 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:04Z","lastTransitionTime":"2026-02-20T08:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.097915 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.097955 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.097997 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.098018 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.098035 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.200292 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.200328 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.200344 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.200365 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.200380 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.302808 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.302875 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.302885 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.302909 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.302921 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.405793 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.405845 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.405864 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.405893 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.405913 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.508047 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.508086 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.508096 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.508113 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.508128 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.611256 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.611301 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.611314 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.611334 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.611346 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.713609 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.713728 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.713756 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.713790 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.713818 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.817063 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.817142 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.817168 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.817201 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.817231 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.921366 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.921426 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.921445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.921470 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.921487 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:05Z","lastTransitionTime":"2026-02-20T08:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:05 crc kubenswrapper[4948]: I0220 08:07:05.995464 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 00:06:08.18944778 +0000 UTC Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.024876 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.024936 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.024950 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.024988 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.025004 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.127700 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.127755 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.127777 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.127808 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.127835 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.231551 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.231614 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.231633 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.231657 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.231677 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.335063 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.335129 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.335146 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.335169 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.335187 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.438474 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.438544 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.438562 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.438586 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.438603 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.541588 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.541651 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.541673 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.541703 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.541725 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.645363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.645423 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.645445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.645477 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.645498 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.721851 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:06 crc kubenswrapper[4948]: E0220 08:07:06.722139 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.722556 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:06 crc kubenswrapper[4948]: E0220 08:07:06.722710 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.723058 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:06 crc kubenswrapper[4948]: E0220 08:07:06.723217 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.725069 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:06 crc kubenswrapper[4948]: E0220 08:07:06.725247 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.749014 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.749054 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.749070 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.749096 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.749110 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.852576 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.852613 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.852625 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.852641 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.852652 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.955507 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.955577 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.955594 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.955612 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.955624 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:06Z","lastTransitionTime":"2026-02-20T08:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:06 crc kubenswrapper[4948]: I0220 08:07:06.996245 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 13:57:31.868688962 +0000 UTC Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.058595 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.058669 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.058707 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.058738 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.058761 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.161639 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.161754 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.161793 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.161828 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.161857 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.264821 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.264896 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.264915 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.264942 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.264961 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.368618 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.368690 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.368708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.368733 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.368752 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.476926 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.477044 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.477070 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.477100 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.477122 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.580685 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.580757 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.580774 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.580801 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.580820 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.684552 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.684619 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.684642 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.684671 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.684694 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.788072 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.788159 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.788174 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.788202 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.788221 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.891106 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.891180 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.891199 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.891231 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.891250 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.995113 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.995193 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.995217 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.995252 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.995277 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:07Z","lastTransitionTime":"2026-02-20T08:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:07 crc kubenswrapper[4948]: I0220 08:07:07.996471 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 06:00:40.051889535 +0000 UTC Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.097914 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.098017 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.098043 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.098074 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.098098 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.201556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.201601 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.201615 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.201635 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.201651 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.303902 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.303957 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.304018 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.304043 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.304059 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.407487 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.407552 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.407570 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.407597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.407615 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.511068 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.511502 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.511706 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.511898 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.512096 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.615121 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.616069 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.616494 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.616925 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.617193 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.720365 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.720427 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.720445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.720473 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.720495 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.721676 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.721737 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.721742 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:08 crc kubenswrapper[4948]: E0220 08:07:08.721830 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:08 crc kubenswrapper[4948]: E0220 08:07:08.722126 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:08 crc kubenswrapper[4948]: E0220 08:07:08.722150 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.721692 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:08 crc kubenswrapper[4948]: E0220 08:07:08.722676 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.823749 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.823816 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.823852 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.823891 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.823916 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.927326 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.927375 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.927392 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.927414 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.927432 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:08Z","lastTransitionTime":"2026-02-20T08:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:08 crc kubenswrapper[4948]: I0220 08:07:08.997557 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 00:21:41.673377034 +0000 UTC Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.030808 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.030884 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.030912 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.030944 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.030968 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.133490 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.133549 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.133566 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.133591 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.133609 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.236635 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.236702 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.236721 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.236748 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.236767 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.338832 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.338888 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.338904 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.338927 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.338944 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.441960 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.442062 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.442087 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.442119 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.442139 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.545184 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.545239 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.545251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.545275 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.545296 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.648390 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.648427 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.648436 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.648450 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.648461 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.744456 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.744494 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.744502 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.744514 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.744524 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: E0220 08:07:09.765393 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:09Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.770470 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.770532 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.770553 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.770577 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.770596 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: E0220 08:07:09.788696 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:09Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.792182 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.792225 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.792236 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.792253 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.792269 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: E0220 08:07:09.809631 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:09Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.814414 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.814458 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.814470 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.814488 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.814501 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: E0220 08:07:09.826846 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:09Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.830199 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.830235 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.830243 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.830258 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.830270 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: E0220 08:07:09.862490 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:09Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:09 crc kubenswrapper[4948]: E0220 08:07:09.862705 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.867251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.867302 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.867313 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.867328 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.867340 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.970059 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.970097 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.970106 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.970119 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.970128 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:09Z","lastTransitionTime":"2026-02-20T08:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:09 crc kubenswrapper[4948]: I0220 08:07:09.998031 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 02:40:18.954398867 +0000 UTC Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.073378 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.073432 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.073449 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.073473 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.073491 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.176997 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.177042 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.177056 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.177074 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.177086 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.280101 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.280165 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.280182 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.280207 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.280224 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.382888 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.382934 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.382946 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.382965 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.382996 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.485900 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.486033 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.486067 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.486099 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.486124 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.589320 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.589384 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.589401 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.589425 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.589443 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.692284 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.692377 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.692397 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.692423 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.692446 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.721883 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.721883 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.722000 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.722064 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:10 crc kubenswrapper[4948]: E0220 08:07:10.722253 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:10 crc kubenswrapper[4948]: E0220 08:07:10.722485 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:10 crc kubenswrapper[4948]: E0220 08:07:10.722570 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:10 crc kubenswrapper[4948]: E0220 08:07:10.722668 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.795590 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.795661 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.795754 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.795785 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.795807 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.898673 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.898722 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.898737 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.898760 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.898777 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:10Z","lastTransitionTime":"2026-02-20T08:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:10 crc kubenswrapper[4948]: I0220 08:07:10.998207 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 05:08:17.046734525 +0000 UTC Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.002092 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.002158 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.002181 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.002214 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.002236 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.105363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.105427 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.105444 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.105471 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.105490 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.209362 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.209428 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.209451 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.209479 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.209497 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.312776 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.313270 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.313440 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.313634 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.313792 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.416735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.416800 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.416819 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.416845 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.416864 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.520133 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.520220 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.520237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.520262 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.520280 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.624091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.624158 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.624176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.624200 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.624220 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.728166 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.728232 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.728249 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.728276 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.728336 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.739755 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ad7145-074b-49df-a0e9-cfec6e3974ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.765039 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.780715 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.808450 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.827643 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.831578 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.831645 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.831667 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.831698 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.831721 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.864591 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.891499 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.918199 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.934624 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.934714 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.934735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.934759 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.934810 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:11Z","lastTransitionTime":"2026-02-20T08:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.937220 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.955069 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.978203 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:11 crc kubenswrapper[4948]: I0220 08:07:11.998458 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 11:11:43.074352609 +0000 UTC Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.000174 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:11Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.020893 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:12Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.038556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.038607 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.038624 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.038650 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.038668 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.048283 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"2026-02-20T08:06:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45\\\\n2026-02-20T08:06:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45 to /host/opt/cni/bin/\\\\n2026-02-20T08:06:17Z [verbose] multus-daemon started\\\\n2026-02-20T08:06:17Z [verbose] Readiness Indicator file check\\\\n2026-02-20T08:07:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:12Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.081767 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:12Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.099284 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:12Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.113341 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:12Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.128061 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:12Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.141532 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:12Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.141772 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.141935 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.142233 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.142478 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.146022 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.248776 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.249072 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.249081 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.249095 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.249105 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.352356 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.352414 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.352428 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.352629 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.352641 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.456632 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.456697 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.456719 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.456746 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.456766 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.566713 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.566788 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.566810 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.566840 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.566864 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.670149 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.670224 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.670246 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.670278 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.670301 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.722339 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.722339 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.722364 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.722504 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:12 crc kubenswrapper[4948]: E0220 08:07:12.723280 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:12 crc kubenswrapper[4948]: E0220 08:07:12.723467 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:12 crc kubenswrapper[4948]: E0220 08:07:12.723652 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:12 crc kubenswrapper[4948]: E0220 08:07:12.723175 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.772595 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.773324 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.773675 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.773859 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.774030 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.877107 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.877182 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.877204 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.877236 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.877254 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.981924 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.982453 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.982603 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.982757 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.982885 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:12Z","lastTransitionTime":"2026-02-20T08:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:12 crc kubenswrapper[4948]: I0220 08:07:12.999713 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 13:33:41.867727325 +0000 UTC Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.086743 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.086781 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.086792 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.086813 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.086824 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.189943 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.190404 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.190520 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.190637 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.190786 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.294275 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.294327 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.294341 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.294361 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.294373 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.398241 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.398314 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.398330 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.398363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.398383 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.500810 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.500852 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.500865 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.500886 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.500900 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.604206 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.604279 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.604297 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.604325 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.604347 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.707736 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.708207 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.708420 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.708590 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.708729 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.722655 4948 scope.go:117] "RemoveContainer" containerID="bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.811254 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.811296 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.811311 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.811333 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.811348 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.915187 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.915243 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.915257 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.915280 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:13 crc kubenswrapper[4948]: I0220 08:07:13.915294 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:13Z","lastTransitionTime":"2026-02-20T08:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.000653 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 18:00:13.65912416 +0000 UTC Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.018959 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.019022 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.019032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.019049 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.019059 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.122018 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.122084 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.122100 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.122125 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.122143 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.224424 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.224459 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.224468 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.224482 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.224491 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.312281 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/2.log" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.315477 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.316350 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.327803 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.327894 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.327921 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.327956 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.328036 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.339659 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.356592 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.370180 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.394662 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.407159 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ad7145-074b-49df-a0e9-cfec6e3974ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.422032 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.431056 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.431086 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.431096 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.431110 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.431120 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.437517 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.451512 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.465423 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.477552 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.490217 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.504746 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"2026-02-20T08:06:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45\\\\n2026-02-20T08:06:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45 to /host/opt/cni/bin/\\\\n2026-02-20T08:06:17Z [verbose] multus-daemon started\\\\n2026-02-20T08:06:17Z [verbose] Readiness Indicator file check\\\\n2026-02-20T08:07:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.520098 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.532231 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.533929 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.533956 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.533980 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.534036 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.534051 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.548021 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.561532 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.575929 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.603114 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.618547 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:14Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.635817 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.635844 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.635853 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.635866 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.635875 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.722375 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.722371 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:14 crc kubenswrapper[4948]: E0220 08:07:14.722562 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.722411 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:14 crc kubenswrapper[4948]: E0220 08:07:14.722721 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.722371 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:14 crc kubenswrapper[4948]: E0220 08:07:14.722842 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:14 crc kubenswrapper[4948]: E0220 08:07:14.722900 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.738495 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.738544 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.738560 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.738578 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.738595 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.841866 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.841927 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.841948 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.842005 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.842027 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.945632 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.945701 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.945719 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.945744 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:14 crc kubenswrapper[4948]: I0220 08:07:14.945764 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:14Z","lastTransitionTime":"2026-02-20T08:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.001407 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 21:27:53.104968457 +0000 UTC Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.048674 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.048749 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.048777 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.048807 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.048828 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.150950 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.151029 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.151042 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.151065 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.151080 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.253483 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.253517 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.253527 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.253545 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.253555 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.322443 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/3.log" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.323720 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/2.log" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.327965 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" exitCode=1 Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.328072 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.328185 4948 scope.go:117] "RemoveContainer" containerID="bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.329370 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:07:15 crc kubenswrapper[4948]: E0220 08:07:15.329730 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.349617 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.357483 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.357556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.357574 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.357602 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.357620 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.370279 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.390297 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.427300 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.451627 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.460571 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.460622 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.460639 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.460664 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.460686 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.469553 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.485185 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.496872 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.519695 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bab0c9c6eaea3116e7b3a8cd9e1046410349efbe66c28eb25a044f13eeb630bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:06:44Z\\\",\\\"message\\\":\\\"ate admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:06:44Z is after 2025-08-24T17:21:41Z]\\\\nI0220 08:06:44.892233 6647 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/community-operators]} name:Service_openshift-marketplace/community-operators_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.189:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d389393c-7ba9-422c-b3f5-06e391d537d2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:14Z\\\",\\\"message\\\":\\\" (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0220 08:07:14.607038 7089 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.607659 7089 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.607737 7089 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:07:14.607794 7089 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.608304 7089 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:07:14.608367 7089 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:07:14.608418 7089 factory.go:656] Stopping watch factory\\\\nI0220 08:07:14.608455 7089 ovnkube.go:599] Stopped ovnkube\\\\nI0220 08:07:14.608456 7089 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.532708 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ad7145-074b-49df-a0e9-cfec6e3974ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.546402 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.564524 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.564584 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.564598 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.564618 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.564635 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.564992 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.579813 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.596609 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.611025 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.626144 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.641477 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"2026-02-20T08:06:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45\\\\n2026-02-20T08:06:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45 to /host/opt/cni/bin/\\\\n2026-02-20T08:06:17Z [verbose] multus-daemon started\\\\n2026-02-20T08:06:17Z [verbose] Readiness Indicator file check\\\\n2026-02-20T08:07:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.653913 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.668376 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.668408 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.668416 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.668431 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.668442 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.668459 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:15Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.770172 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.770210 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.770221 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.770240 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.770251 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.873110 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.873165 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.873178 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.873197 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.873209 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.976174 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.976240 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.976253 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.976281 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:15 crc kubenswrapper[4948]: I0220 08:07:15.976300 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:15Z","lastTransitionTime":"2026-02-20T08:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:16 crc kubenswrapper[4948]: I0220 08:07:16.001856 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 06:39:37.670922273 +0000 UTC Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:16.520379 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:16.520436 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.520576 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.520642 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:08:20.520621404 +0000 UTC m=+149.495116224 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.520844 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.520869 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 08:08:20.52086201 +0000 UTC m=+149.495356820 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:16.621404 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:16.621472 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:16.621520 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621630 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621645 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621658 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621694 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 08:08:20.621681948 +0000 UTC m=+149.596176768 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621872 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:20.621863943 +0000 UTC m=+149.596358763 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621923 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621934 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621942 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:16.621991 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 08:08:20.621983595 +0000 UTC m=+149.596478415 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.002356 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 04:32:58.656160035 +0000 UTC Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.282704 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:17.282850 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.282854 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.282881 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.283128 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:17.282947 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:17.283477 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:17.283860 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.285536 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.285573 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.285585 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.285604 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.285618 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:17Z","lastTransitionTime":"2026-02-20T08:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.289050 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/3.log" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.293252 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:07:17 crc kubenswrapper[4948]: E0220 08:07:17.293886 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.311941 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.328096 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.347424 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.364076 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.379556 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.388073 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.388119 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.388128 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.388144 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.388157 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:17Z","lastTransitionTime":"2026-02-20T08:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.398861 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.412584 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.430621 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"2026-02-20T08:06:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45\\\\n2026-02-20T08:06:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45 to /host/opt/cni/bin/\\\\n2026-02-20T08:06:17Z [verbose] multus-daemon started\\\\n2026-02-20T08:06:17Z [verbose] Readiness Indicator file check\\\\n2026-02-20T08:07:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.443564 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.466946 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.482187 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.491265 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.491337 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.491359 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.491391 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.491412 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:17Z","lastTransitionTime":"2026-02-20T08:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.497547 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.515041 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.531689 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.554760 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:14Z\\\",\\\"message\\\":\\\" (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0220 08:07:14.607038 7089 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.607659 7089 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.607737 7089 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:07:14.607794 7089 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.608304 7089 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:07:14.608367 7089 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:07:14.608418 7089 factory.go:656] Stopping watch factory\\\\nI0220 08:07:14.608455 7089 ovnkube.go:599] Stopped ovnkube\\\\nI0220 08:07:14.608456 7089 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:07:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.569598 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ad7145-074b-49df-a0e9-cfec6e3974ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.587621 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.593830 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.593893 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.593913 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.593938 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.593958 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:17Z","lastTransitionTime":"2026-02-20T08:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.606965 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.627021 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:17Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.697024 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.697087 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.697107 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.697130 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.697145 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:17Z","lastTransitionTime":"2026-02-20T08:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.800106 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.800403 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.800473 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.800537 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.800596 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:17Z","lastTransitionTime":"2026-02-20T08:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.904109 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.904175 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.904198 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.904228 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:17 crc kubenswrapper[4948]: I0220 08:07:17.904251 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:17Z","lastTransitionTime":"2026-02-20T08:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.003280 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 11:08:55.305869874 +0000 UTC Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.008403 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.008462 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.008479 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.008505 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.008522 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.111604 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.111669 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.111690 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.111718 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.111741 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.214915 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.215028 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.215066 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.215097 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.215117 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.318431 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.318484 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.318498 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.318518 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.318532 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.422310 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.422652 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.422666 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.422682 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.422693 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.525229 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.525299 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.525317 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.525342 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.525361 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.628820 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.629056 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.629085 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.629120 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.629144 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.722426 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.722498 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.722424 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:18 crc kubenswrapper[4948]: E0220 08:07:18.722686 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:18 crc kubenswrapper[4948]: E0220 08:07:18.722791 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:18 crc kubenswrapper[4948]: E0220 08:07:18.722891 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.731698 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.731765 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.731787 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.731815 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.731840 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.835804 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.835862 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.835879 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.835901 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.835918 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.939032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.939091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.939113 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.939140 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:18 crc kubenswrapper[4948]: I0220 08:07:18.939163 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:18Z","lastTransitionTime":"2026-02-20T08:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.004556 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 01:32:47.679988356 +0000 UTC Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.041486 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.041609 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.041628 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.041653 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.041671 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.144544 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.144612 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.144635 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.144659 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.144677 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.248090 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.248174 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.248205 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.248248 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.248277 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.351029 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.351091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.351114 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.351144 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.351166 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.454125 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.454184 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.454202 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.454237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.454260 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.557886 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.558008 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.558032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.558061 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.558082 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.660930 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.661008 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.661028 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.661055 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.661074 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.721577 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:19 crc kubenswrapper[4948]: E0220 08:07:19.721771 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.763837 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.763894 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.763912 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.763934 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.763951 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.866763 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.866815 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.866832 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.866853 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.866870 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.969936 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.970021 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.970046 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.970072 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:19 crc kubenswrapper[4948]: I0220 08:07:19.970096 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:19Z","lastTransitionTime":"2026-02-20T08:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.004871 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 07:59:29.364927949 +0000 UTC Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.073631 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.073701 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.073728 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.073758 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.073782 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.083131 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.083395 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.083441 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.083478 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.083496 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.105578 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.110568 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.110629 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.110720 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.110751 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.110768 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.132568 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.142525 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.143119 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.143151 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.143181 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.143202 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.165532 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.171558 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.171622 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.171641 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.171668 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.171685 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.193911 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.198702 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.198756 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.198774 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.198800 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.198820 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.220357 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f36346c-e92f-4a00-a0af-e0652f71277d\\\",\\\"systemUUID\\\":\\\"51d89745-b91f-4e85-9724-1ef53e3862a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:20Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.220644 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.223002 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.223069 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.223090 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.223113 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.223130 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.326882 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.326948 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.327014 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.327048 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.327071 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.429731 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.429787 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.429811 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.429839 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.429860 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.536112 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.536160 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.536176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.536194 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.536205 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.639927 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.640024 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.640046 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.640072 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.640089 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.721773 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.721847 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.721773 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.722018 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.722164 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:20 crc kubenswrapper[4948]: E0220 08:07:20.722303 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.742697 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.742752 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.742790 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.742821 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.742844 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.845555 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.845614 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.845637 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.845660 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.845681 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.948848 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.948902 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.948912 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.948928 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:20 crc kubenswrapper[4948]: I0220 08:07:20.948942 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:20Z","lastTransitionTime":"2026-02-20T08:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.006458 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 23:15:07.587189321 +0000 UTC Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.052377 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.052426 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.052445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.052467 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.052515 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.156905 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.156952 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.156961 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.157003 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.157016 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.259757 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.259801 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.259813 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.259909 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.259927 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.363406 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.363447 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.363456 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.363472 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.363485 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.465724 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.465752 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.465761 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.465774 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.465783 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.567944 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.567992 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.568001 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.568020 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.568031 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.671858 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.671928 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.671941 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.671964 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.671999 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.722684 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:21 crc kubenswrapper[4948]: E0220 08:07:21.722952 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.746097 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bc8c9ea-350e-45e5-af21-f268fb6cbffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"ed a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\"\\\\nI0220 08:06:11.996929 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 08:06:12.001143 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 08:06:12.001168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 08:06:12.001192 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 08:06:12.001197 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 08:06:12.009746 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 08:06:12.009780 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009787 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 08:06:12.009793 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 08:06:12.009797 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 08:06:12.009800 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 08:06:12.009803 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 08:06:12.010029 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0220 08:06:12.015892 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2986946918/tls.crt::/tmp/serving-cert-2986946918/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1771574755\\\\\\\\\\\\\\\" (2026-02-20 08:05:54 +0000 UTC to 2026-03-22 08:05:55 +0000 UTC (now=2026-02-20 08:06:12.015833387 +0000 UTC))\\\\\\\"\\\\nF0220 08:06:12.015947 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.765608 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae14aa1-2ba5-4630-b0b9-97cef2fece9c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://001bf218b5793a1101d9960b6d81173d67c5f6259c21d780623dad9808adfd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5f1450fc3f715e0082ad316af3f03ca0d2d8f32b5efa9850e0694864f79153e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a52eab83c2f040b6bd090f63f55a7da942f1b9a219ae29f14fc2b9930b07da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://673c2dfbe564d53841742860ac3e77498ddb123b1e1f53dbf2a9d67399e04054\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.774264 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.774312 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.774327 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.774352 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.774367 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.779237 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07a26aa870472d7f0b52c95662777df37a8aa2dedbf9f613156b0dbc33d76a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a584a8a79e4507a83113203ced4b9aa58a09afd291672cd788e03f2a925e0237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.795000 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.812402 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0449c8ab-b11e-428d-95f5-2f9588e3ca35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35af14834dfe83bb386c833077e8dec1f3c902203420abdfc6e85a61814c5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78953bdfd5717bd7fb3d98acbc04963e8868c0b8b5fbfa5e3e029eb0d9aa6378\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45d15dafc529ce49a5db7681d4278a1a33374be31a388195228847f793e908ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.826372 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.844483 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f92c4de77ec924cae5f5978ecb0f8f458eee13eb34f6b70b0114f9ae9f008d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.864132 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7s56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b96124e4-0a74-4578-9142-fd728eb9f99e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:02Z\\\",\\\"message\\\":\\\"2026-02-20T08:06:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45\\\\n2026-02-20T08:06:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b42b7938-8525-4811-9d9f-355344db9d45 to /host/opt/cni/bin/\\\\n2026-02-20T08:06:17Z [verbose] multus-daemon started\\\\n2026-02-20T08:06:17Z [verbose] Readiness Indicator file check\\\\n2026-02-20T08:07:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2fpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7s56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.876634 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.876674 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.876686 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.876702 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.876715 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.881407 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xd86g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xd86g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.902329 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7767158a-4a6f-4c72-8112-3549613da1df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c41603802102d34d3ae52828855f0626ee97fe541ba55af3d3dcb00f1a06ca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8f8039ba22fb5059c16ab3c3938aa95370036c1cf0a48e6d2db4f123541c22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54442906083c1d59577e91dce6c0d65e765231523b156de107d4231a4379c7dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c9c9106da27bc6f96be49e6ba100be2dc8a09034ca21e30e8ac015b48a7fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a7cdf341f90dbda3d95e3df0f00ba9ccbd033c6812395d6d30dc3720eeee154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb338cd6c6037248bc567ce5c2785c85d1e1e5ee49407eee7b2758be98b6cfe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b13269f6583f3d7d9f2eda1cc03f36e7d286ec928d04b7e6441db3bd723bbfc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f878e6d27db45bc18f547d6f90bca7d693d0411da2a8faebd313fe431c16aeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.915541 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.928083 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-nsk7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba900a81-fc7b-40b3-942c-097c47812bf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eeeaecaef0d9e47eac4c1f764778b88ffa8105d6a5a8c0788a3b784ee01802c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dc562\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-nsk7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.942271 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b664cac-f1be-4b54-b85d-fede8b4d7d31\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6cfa24bc2ee3daae226cdfb1ec830d8a97c86eed37ca8ce2d0c10878627bcc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cc8878e457802d4970c4db52fc332cf96d26ab80c90cbe1c3160b8c02b385e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qhj9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-j8xql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.955130 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f75fc696f88e6c634f8f8c285835de7826d021ace167069f3d61774a0fbee98e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnmlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h4ww2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.973719 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a28bd7-2b83-43f8-b803-bfe41516e071\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T08:07:14Z\\\",\\\"message\\\":\\\" (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0220 08:07:14.607038 7089 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.607659 7089 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.607737 7089 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 08:07:14.607794 7089 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0220 08:07:14.608304 7089 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 08:07:14.608367 7089 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 08:07:14.608418 7089 factory.go:656] Stopping watch factory\\\\nI0220 08:07:14.608455 7089 ovnkube.go:599] Stopped ovnkube\\\\nI0220 08:07:14.608456 7089 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T08:07:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtrs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7s8gg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.978890 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.978922 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.978934 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.978952 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.978963 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:21Z","lastTransitionTime":"2026-02-20T08:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:21 crc kubenswrapper[4948]: I0220 08:07:21.988587 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ad7145-074b-49df-a0e9-cfec6e3974ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67b60ad2efa00639a153f39923bfdba3b40b6b3f5126207b4f0e05abdf9f43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:05:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3bf598724d4df70deb63c98b0836455d7f1f3d83fd0cac39958974e4365221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:05:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:05:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:05:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:21Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.006839 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 15:28:44.025217868 +0000 UTC Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.007312 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a322c7e6c53b414e88899b65f5f714441e36b0b941f53709ac9875ecc0d8ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.020550 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z8rrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9cb414-e3c0-4e40-92b1-42f05565e44c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://180eead71e3d8b6a54ee83ac9c9a39be2ea362f006e0695f36b7a901a8c7725b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tzrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z8rrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.038524 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8frp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5757f611-4b0c-49dc-a076-c90ebaad78d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T08:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78b4411ee2fa1736d0f9332860c74322784f7471c2683c8bfe2673ac93401b5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T08:06:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dface5f477c248b5207a61c87eb507fe151dffa9b6ed35c8eb9ea20088898069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9359881f5ce04f7f53c2718515595553067881ed45d4ab2a1e30bc48f9319b32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0e453f489cf143c0ec47592e5c71730282cd7b0c0b51741d2165210b48f5bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b282884a0b925cd25588a226893b87f6c54810bd15a6f4a990cdd4487c84c155\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca260a43deb1f3903961cf3cb68fb9908183433954bc62e891528cd89aed0cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a0587d2a33fa2dbdfc9bbffdb7591218612d30ab8514e45a8ec68a84f08ac4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T08:06:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T08:06:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9689\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T08:06:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8frp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T08:07:22Z is after 2025-08-24T17:21:41Z" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.082032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.082091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.082102 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.082121 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.082134 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.185067 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.185151 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.185175 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.185207 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.185233 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.287862 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.287942 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.288025 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.288080 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.288104 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.391176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.391251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.391270 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.391294 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.391313 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.494629 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.494690 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.494707 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.494732 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.494750 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.598051 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.598109 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.598145 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.598169 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.598187 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.701073 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.701132 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.701152 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.701176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.701196 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.721818 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.721944 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:22 crc kubenswrapper[4948]: E0220 08:07:22.722036 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.721944 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:22 crc kubenswrapper[4948]: E0220 08:07:22.722202 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:22 crc kubenswrapper[4948]: E0220 08:07:22.722309 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.804817 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.804879 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.804900 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.804926 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.804947 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.908448 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.908489 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.908500 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.908515 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:22 crc kubenswrapper[4948]: I0220 08:07:22.908527 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:22Z","lastTransitionTime":"2026-02-20T08:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.008005 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 13:51:16.041023369 +0000 UTC Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.011707 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.011771 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.011789 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.011817 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.011836 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.115132 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.115197 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.115221 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.115251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.115276 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.219147 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.219236 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.219264 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.219298 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.219323 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.321756 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.321833 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.321859 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.321891 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.321912 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.444116 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.444210 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.444228 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.444259 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.444283 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.548780 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.548861 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.548888 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.548935 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.548960 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.652595 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.652851 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.653029 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.653236 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.653420 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.722524 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:23 crc kubenswrapper[4948]: E0220 08:07:23.722879 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.756827 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.756926 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.756945 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.756994 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.757009 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.860730 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.860930 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.861196 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.861393 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.861547 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.964820 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.964893 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.964922 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.964958 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:23 crc kubenswrapper[4948]: I0220 08:07:23.965017 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:23Z","lastTransitionTime":"2026-02-20T08:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.008914 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 13:55:27.866622145 +0000 UTC Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.069127 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.069193 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.069216 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.069245 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.069265 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.172533 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.172599 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.172625 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.172652 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.172671 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.274551 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.274604 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.274621 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.274644 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.274660 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.377527 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.377586 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.377611 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.377646 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.377671 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.480469 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.480526 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.480543 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.480570 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.480589 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.584518 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.585544 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.585769 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.585942 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.586131 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.722737 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.723165 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:24 crc kubenswrapper[4948]: E0220 08:07:24.723316 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.723405 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:24 crc kubenswrapper[4948]: E0220 08:07:24.723498 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:24 crc kubenswrapper[4948]: E0220 08:07:24.723644 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.723699 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.723758 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.723778 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.723801 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.723819 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.826173 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.826254 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.826273 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.826305 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.826394 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.929923 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.930038 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.930056 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.930081 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:24 crc kubenswrapper[4948]: I0220 08:07:24.930102 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:24Z","lastTransitionTime":"2026-02-20T08:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.010053 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 23:00:32.684727381 +0000 UTC Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.032744 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.032793 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.032808 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.032831 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.032846 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.135885 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.136002 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.136023 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.136048 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.136067 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.239253 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.239317 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.239334 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.239362 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.239380 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.342483 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.342538 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.342557 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.342583 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.342601 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.445040 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.445101 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.445143 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.445171 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.445190 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.547512 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.547548 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.547556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.547569 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.547579 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.650624 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.650665 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.650704 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.650722 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.650735 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.721843 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:25 crc kubenswrapper[4948]: E0220 08:07:25.722012 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.752620 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.753482 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.753543 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.753580 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.753604 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.856883 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.856945 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.856967 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.857020 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.857043 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.959365 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.959437 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.959461 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.959493 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:25 crc kubenswrapper[4948]: I0220 08:07:25.959517 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:25Z","lastTransitionTime":"2026-02-20T08:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.010210 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 18:22:37.494920974 +0000 UTC Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.061836 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.061867 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.061876 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.061889 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.061897 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.164348 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.164452 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.164525 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.164557 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.164580 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.267176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.267509 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.267619 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.267726 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.267825 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.371204 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.371270 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.371288 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.371315 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.371346 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.474735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.475115 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.475210 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.475291 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.475359 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.578707 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.579079 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.579233 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.579383 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.579523 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.682760 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.683607 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.683747 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.683880 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.684097 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.722149 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.722198 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:26 crc kubenswrapper[4948]: E0220 08:07:26.722389 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:26 crc kubenswrapper[4948]: E0220 08:07:26.722524 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.722717 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:26 crc kubenswrapper[4948]: E0220 08:07:26.723098 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.786754 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.786791 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.786802 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.786816 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.786827 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.889040 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.889093 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.889105 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.889119 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.889131 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.992032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.992112 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.992138 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.992167 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:26 crc kubenswrapper[4948]: I0220 08:07:26.992186 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:26Z","lastTransitionTime":"2026-02-20T08:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.011393 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 12:04:10.677207005 +0000 UTC Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.095634 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.096045 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.096077 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.096110 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.096135 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.199403 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.199454 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.199466 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.199486 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.199504 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.303033 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.303121 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.303140 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.303168 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.303187 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.406599 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.406949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.407248 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.407455 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.407647 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.511269 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.511631 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.511786 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.511934 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.512170 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.617613 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.618016 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.618244 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.618431 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.618634 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.722862 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:27 crc kubenswrapper[4948]: E0220 08:07:27.724003 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.724376 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.724499 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.724582 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.724671 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.724761 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.726311 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:07:27 crc kubenswrapper[4948]: E0220 08:07:27.726606 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.828044 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.828120 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.828138 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.828167 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.828188 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.931828 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.932269 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.932495 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.932698 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:27 crc kubenswrapper[4948]: I0220 08:07:27.932859 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:27Z","lastTransitionTime":"2026-02-20T08:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.012020 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 02:33:57.306471255 +0000 UTC Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.036324 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.036387 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.036402 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.036421 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.036437 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.140100 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.140511 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.140681 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.140845 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.141033 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.244342 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.244414 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.244454 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.244492 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.244520 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.347597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.347665 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.347688 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.347722 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.347743 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.451582 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.452291 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.452537 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.452753 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.452925 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.555671 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.556115 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.556380 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.556597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.556808 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.659458 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.659724 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.659795 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.660049 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.660129 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.721985 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.722094 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.722046 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:28 crc kubenswrapper[4948]: E0220 08:07:28.722402 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:28 crc kubenswrapper[4948]: E0220 08:07:28.722561 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:28 crc kubenswrapper[4948]: E0220 08:07:28.722694 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.762251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.762281 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.762290 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.762309 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.762319 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.865162 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.865587 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.865609 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.865634 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.865651 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.968564 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.968616 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.968628 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.968645 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:28 crc kubenswrapper[4948]: I0220 08:07:28.968659 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:28Z","lastTransitionTime":"2026-02-20T08:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.012855 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 08:38:50.010146446 +0000 UTC Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.072239 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.072281 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.072292 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.072309 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.072321 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.175541 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.175621 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.175639 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.175668 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.175690 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.279023 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.279092 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.279116 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.279142 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.279165 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.382108 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.382168 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.382185 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.382207 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.382226 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.485138 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.485206 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.485223 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.485252 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.485271 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.587834 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.587892 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.587909 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.587932 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.587949 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.690841 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.690900 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.690910 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.690928 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.690941 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.722417 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:29 crc kubenswrapper[4948]: E0220 08:07:29.722965 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.794300 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.794345 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.794357 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.794375 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.794387 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.896844 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.896911 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.896930 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.896954 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:29 crc kubenswrapper[4948]: I0220 08:07:29.897002 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:29Z","lastTransitionTime":"2026-02-20T08:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.000363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.000425 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.000443 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.000466 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.000486 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:30Z","lastTransitionTime":"2026-02-20T08:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.013786 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 07:03:19.413260893 +0000 UTC Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.104024 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.104084 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.104101 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.104122 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.104136 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:30Z","lastTransitionTime":"2026-02-20T08:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.206529 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.206569 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.206581 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.206598 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.206612 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:30Z","lastTransitionTime":"2026-02-20T08:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.310257 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.310314 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.310330 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.310355 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.310372 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:30Z","lastTransitionTime":"2026-02-20T08:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.387942 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.388033 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.388052 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.388080 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.388102 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T08:07:30Z","lastTransitionTime":"2026-02-20T08:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.447898 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64"] Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.448377 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.452017 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.453364 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.459705 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.459734 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.484688 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=77.484657787 podStartE2EDuration="1m17.484657787s" podCreationTimestamp="2026-02-20 08:06:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.484400241 +0000 UTC m=+99.458895081" watchObservedRunningTime="2026-02-20 08:07:30.484657787 +0000 UTC m=+99.459152637" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.486960 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1e7d2472-cdda-4f5b-996f-547bb51497d8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.487107 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1e7d2472-cdda-4f5b-996f-547bb51497d8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.487169 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1e7d2472-cdda-4f5b-996f-547bb51497d8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.487259 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1e7d2472-cdda-4f5b-996f-547bb51497d8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.487304 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e7d2472-cdda-4f5b-996f-547bb51497d8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.532262 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-x7s56" podStartSLOduration=76.532237523 podStartE2EDuration="1m16.532237523s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.532015438 +0000 UTC m=+99.506510258" watchObservedRunningTime="2026-02-20 08:07:30.532237523 +0000 UTC m=+99.506732383" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.588080 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1e7d2472-cdda-4f5b-996f-547bb51497d8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.588128 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1e7d2472-cdda-4f5b-996f-547bb51497d8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.588088 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=78.588053027 podStartE2EDuration="1m18.588053027s" podCreationTimestamp="2026-02-20 08:06:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.56741853 +0000 UTC m=+99.541913350" watchObservedRunningTime="2026-02-20 08:07:30.588053027 +0000 UTC m=+99.562547887" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.588879 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1e7d2472-cdda-4f5b-996f-547bb51497d8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.589126 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1e7d2472-cdda-4f5b-996f-547bb51497d8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.588162 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1e7d2472-cdda-4f5b-996f-547bb51497d8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.589278 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e7d2472-cdda-4f5b-996f-547bb51497d8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.589327 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1e7d2472-cdda-4f5b-996f-547bb51497d8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.589367 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1e7d2472-cdda-4f5b-996f-547bb51497d8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.597157 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e7d2472-cdda-4f5b-996f-547bb51497d8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.602029 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-nsk7l" podStartSLOduration=76.602004833 podStartE2EDuration="1m16.602004833s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.601757447 +0000 UTC m=+99.576252267" watchObservedRunningTime="2026-02-20 08:07:30.602004833 +0000 UTC m=+99.576499653" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.615329 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1e7d2472-cdda-4f5b-996f-547bb51497d8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nkq64\" (UID: \"1e7d2472-cdda-4f5b-996f-547bb51497d8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.635480 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-j8xql" podStartSLOduration=76.635459219 podStartE2EDuration="1m16.635459219s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.617414964 +0000 UTC m=+99.591909834" watchObservedRunningTime="2026-02-20 08:07:30.635459219 +0000 UTC m=+99.609954039" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.667494 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=32.667471299 podStartE2EDuration="32.667471299s" podCreationTimestamp="2026-02-20 08:06:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.652210552 +0000 UTC m=+99.626705372" watchObservedRunningTime="2026-02-20 08:07:30.667471299 +0000 UTC m=+99.641966119" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.681735 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-z8rrb" podStartSLOduration=76.681705582 podStartE2EDuration="1m16.681705582s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.681022425 +0000 UTC m=+99.655517245" watchObservedRunningTime="2026-02-20 08:07:30.681705582 +0000 UTC m=+99.656200422" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.703150 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-8frp4" podStartSLOduration=76.703133738 podStartE2EDuration="1m16.703133738s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.701834947 +0000 UTC m=+99.676329777" watchObservedRunningTime="2026-02-20 08:07:30.703133738 +0000 UTC m=+99.677628558" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.716825 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podStartSLOduration=76.716807847 podStartE2EDuration="1m16.716807847s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.716351616 +0000 UTC m=+99.690846436" watchObservedRunningTime="2026-02-20 08:07:30.716807847 +0000 UTC m=+99.691302667" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.721689 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.721727 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.721768 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:30 crc kubenswrapper[4948]: E0220 08:07:30.721800 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:30 crc kubenswrapper[4948]: E0220 08:07:30.721907 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:30 crc kubenswrapper[4948]: E0220 08:07:30.722052 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.774383 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.774363714 podStartE2EDuration="1m18.774363714s" podCreationTimestamp="2026-02-20 08:06:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.773096403 +0000 UTC m=+99.747591223" watchObservedRunningTime="2026-02-20 08:07:30.774363714 +0000 UTC m=+99.748858534" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.775080 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" Feb 20 08:07:30 crc kubenswrapper[4948]: I0220 08:07:30.816368 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=45.816341115 podStartE2EDuration="45.816341115s" podCreationTimestamp="2026-02-20 08:06:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:30.799410067 +0000 UTC m=+99.773904887" watchObservedRunningTime="2026-02-20 08:07:30.816341115 +0000 UTC m=+99.790835935" Feb 20 08:07:31 crc kubenswrapper[4948]: I0220 08:07:31.014256 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 12:54:48.8422828 +0000 UTC Feb 20 08:07:31 crc kubenswrapper[4948]: I0220 08:07:31.014343 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 20 08:07:31 crc kubenswrapper[4948]: I0220 08:07:31.024557 4948 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 20 08:07:31 crc kubenswrapper[4948]: I0220 08:07:31.388994 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" event={"ID":"1e7d2472-cdda-4f5b-996f-547bb51497d8","Type":"ContainerStarted","Data":"a4f04e4cb0bf2b7c7d9aef9c44aae1318cc8087599dd94e6f0ca316bcaec97db"} Feb 20 08:07:31 crc kubenswrapper[4948]: I0220 08:07:31.389082 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" event={"ID":"1e7d2472-cdda-4f5b-996f-547bb51497d8","Type":"ContainerStarted","Data":"fcd683dea6851584660c8afd3237ccb1cbd4919e2f27216d734520806e1173af"} Feb 20 08:07:31 crc kubenswrapper[4948]: I0220 08:07:31.414762 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkq64" podStartSLOduration=77.414738207 podStartE2EDuration="1m17.414738207s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:31.414469371 +0000 UTC m=+100.388964211" watchObservedRunningTime="2026-02-20 08:07:31.414738207 +0000 UTC m=+100.389233047" Feb 20 08:07:31 crc kubenswrapper[4948]: I0220 08:07:31.722141 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:31 crc kubenswrapper[4948]: E0220 08:07:31.724589 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:32 crc kubenswrapper[4948]: I0220 08:07:32.712187 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:32 crc kubenswrapper[4948]: E0220 08:07:32.712447 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:07:32 crc kubenswrapper[4948]: E0220 08:07:32.713766 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs podName:62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd nodeName:}" failed. No retries permitted until 2026-02-20 08:08:36.713414616 +0000 UTC m=+165.687909476 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs") pod "network-metrics-daemon-xd86g" (UID: "62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 08:07:32 crc kubenswrapper[4948]: I0220 08:07:32.722246 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:32 crc kubenswrapper[4948]: I0220 08:07:32.722246 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:32 crc kubenswrapper[4948]: E0220 08:07:32.722453 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:32 crc kubenswrapper[4948]: E0220 08:07:32.722595 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:32 crc kubenswrapper[4948]: I0220 08:07:32.722880 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:32 crc kubenswrapper[4948]: E0220 08:07:32.723347 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:33 crc kubenswrapper[4948]: I0220 08:07:33.722182 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:33 crc kubenswrapper[4948]: E0220 08:07:33.722378 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:34 crc kubenswrapper[4948]: I0220 08:07:34.722375 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:34 crc kubenswrapper[4948]: I0220 08:07:34.722445 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:34 crc kubenswrapper[4948]: I0220 08:07:34.722503 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:34 crc kubenswrapper[4948]: E0220 08:07:34.722615 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:34 crc kubenswrapper[4948]: E0220 08:07:34.723023 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:34 crc kubenswrapper[4948]: E0220 08:07:34.723099 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:35 crc kubenswrapper[4948]: I0220 08:07:35.721714 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:35 crc kubenswrapper[4948]: E0220 08:07:35.721968 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:36 crc kubenswrapper[4948]: I0220 08:07:36.721988 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:36 crc kubenswrapper[4948]: I0220 08:07:36.722010 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:36 crc kubenswrapper[4948]: E0220 08:07:36.722135 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:36 crc kubenswrapper[4948]: I0220 08:07:36.722207 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:36 crc kubenswrapper[4948]: E0220 08:07:36.722314 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:36 crc kubenswrapper[4948]: E0220 08:07:36.722443 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:37 crc kubenswrapper[4948]: I0220 08:07:37.722112 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:37 crc kubenswrapper[4948]: E0220 08:07:37.722339 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:38 crc kubenswrapper[4948]: I0220 08:07:38.721464 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:38 crc kubenswrapper[4948]: I0220 08:07:38.721537 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:38 crc kubenswrapper[4948]: I0220 08:07:38.721482 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:38 crc kubenswrapper[4948]: E0220 08:07:38.721668 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:38 crc kubenswrapper[4948]: E0220 08:07:38.721792 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:38 crc kubenswrapper[4948]: E0220 08:07:38.721937 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:39 crc kubenswrapper[4948]: I0220 08:07:39.721950 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:39 crc kubenswrapper[4948]: E0220 08:07:39.722427 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:40 crc kubenswrapper[4948]: I0220 08:07:40.721432 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:40 crc kubenswrapper[4948]: I0220 08:07:40.721529 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:40 crc kubenswrapper[4948]: I0220 08:07:40.721456 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:40 crc kubenswrapper[4948]: E0220 08:07:40.721783 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:40 crc kubenswrapper[4948]: E0220 08:07:40.721893 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:40 crc kubenswrapper[4948]: E0220 08:07:40.721670 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:41 crc kubenswrapper[4948]: I0220 08:07:41.721957 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:41 crc kubenswrapper[4948]: E0220 08:07:41.725092 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:41 crc kubenswrapper[4948]: I0220 08:07:41.726893 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:07:41 crc kubenswrapper[4948]: E0220 08:07:41.727371 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:07:42 crc kubenswrapper[4948]: I0220 08:07:42.722531 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:42 crc kubenswrapper[4948]: I0220 08:07:42.722540 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:42 crc kubenswrapper[4948]: E0220 08:07:42.722797 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:42 crc kubenswrapper[4948]: E0220 08:07:42.722934 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:42 crc kubenswrapper[4948]: I0220 08:07:42.722540 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:42 crc kubenswrapper[4948]: E0220 08:07:42.724173 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:43 crc kubenswrapper[4948]: I0220 08:07:43.722117 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:43 crc kubenswrapper[4948]: E0220 08:07:43.722578 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:44 crc kubenswrapper[4948]: I0220 08:07:44.721843 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:44 crc kubenswrapper[4948]: I0220 08:07:44.721850 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:44 crc kubenswrapper[4948]: I0220 08:07:44.721915 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:44 crc kubenswrapper[4948]: E0220 08:07:44.722554 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:44 crc kubenswrapper[4948]: E0220 08:07:44.722759 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:44 crc kubenswrapper[4948]: E0220 08:07:44.722935 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:45 crc kubenswrapper[4948]: I0220 08:07:45.721683 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:45 crc kubenswrapper[4948]: E0220 08:07:45.721894 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:46 crc kubenswrapper[4948]: I0220 08:07:46.722327 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:46 crc kubenswrapper[4948]: I0220 08:07:46.722401 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:46 crc kubenswrapper[4948]: I0220 08:07:46.722328 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:46 crc kubenswrapper[4948]: E0220 08:07:46.722546 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:46 crc kubenswrapper[4948]: E0220 08:07:46.722660 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:46 crc kubenswrapper[4948]: E0220 08:07:46.722759 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:47 crc kubenswrapper[4948]: I0220 08:07:47.721854 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:47 crc kubenswrapper[4948]: E0220 08:07:47.722437 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.473572 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/1.log" Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.475249 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/0.log" Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.475495 4948 generic.go:334] "Generic (PLEG): container finished" podID="b96124e4-0a74-4578-9142-fd728eb9f99e" containerID="2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c" exitCode=1 Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.475605 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7s56" event={"ID":"b96124e4-0a74-4578-9142-fd728eb9f99e","Type":"ContainerDied","Data":"2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c"} Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.475892 4948 scope.go:117] "RemoveContainer" containerID="34d766fd566183e0a677be578b208a2dbd78ffc6551509b294a706715424ac5e" Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.476516 4948 scope.go:117] "RemoveContainer" containerID="2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c" Feb 20 08:07:48 crc kubenswrapper[4948]: E0220 08:07:48.476797 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-x7s56_openshift-multus(b96124e4-0a74-4578-9142-fd728eb9f99e)\"" pod="openshift-multus/multus-x7s56" podUID="b96124e4-0a74-4578-9142-fd728eb9f99e" Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.721533 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.721557 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:48 crc kubenswrapper[4948]: E0220 08:07:48.721740 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:48 crc kubenswrapper[4948]: I0220 08:07:48.721563 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:48 crc kubenswrapper[4948]: E0220 08:07:48.721951 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:48 crc kubenswrapper[4948]: E0220 08:07:48.722192 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:49 crc kubenswrapper[4948]: I0220 08:07:49.482542 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/1.log" Feb 20 08:07:49 crc kubenswrapper[4948]: I0220 08:07:49.721794 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:49 crc kubenswrapper[4948]: E0220 08:07:49.722053 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:50 crc kubenswrapper[4948]: I0220 08:07:50.722158 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:50 crc kubenswrapper[4948]: I0220 08:07:50.722280 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:50 crc kubenswrapper[4948]: E0220 08:07:50.722380 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:50 crc kubenswrapper[4948]: I0220 08:07:50.722469 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:50 crc kubenswrapper[4948]: E0220 08:07:50.723299 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:50 crc kubenswrapper[4948]: E0220 08:07:50.723376 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:51 crc kubenswrapper[4948]: E0220 08:07:51.696518 4948 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 20 08:07:51 crc kubenswrapper[4948]: I0220 08:07:51.722020 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:51 crc kubenswrapper[4948]: E0220 08:07:51.723332 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:52 crc kubenswrapper[4948]: E0220 08:07:52.295668 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 20 08:07:52 crc kubenswrapper[4948]: I0220 08:07:52.721469 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:52 crc kubenswrapper[4948]: I0220 08:07:52.721493 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:52 crc kubenswrapper[4948]: E0220 08:07:52.721708 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:52 crc kubenswrapper[4948]: E0220 08:07:52.721799 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:52 crc kubenswrapper[4948]: I0220 08:07:52.721507 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:52 crc kubenswrapper[4948]: E0220 08:07:52.722560 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:53 crc kubenswrapper[4948]: I0220 08:07:53.722695 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:53 crc kubenswrapper[4948]: E0220 08:07:53.723004 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:53 crc kubenswrapper[4948]: I0220 08:07:53.723108 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:07:53 crc kubenswrapper[4948]: E0220 08:07:53.723513 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7s8gg_openshift-ovn-kubernetes(d6a28bd7-2b83-43f8-b803-bfe41516e071)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" Feb 20 08:07:54 crc kubenswrapper[4948]: I0220 08:07:54.722386 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:54 crc kubenswrapper[4948]: I0220 08:07:54.722436 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:54 crc kubenswrapper[4948]: I0220 08:07:54.722397 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:54 crc kubenswrapper[4948]: E0220 08:07:54.722540 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:54 crc kubenswrapper[4948]: E0220 08:07:54.722782 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:54 crc kubenswrapper[4948]: E0220 08:07:54.722880 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:55 crc kubenswrapper[4948]: I0220 08:07:55.722372 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:55 crc kubenswrapper[4948]: E0220 08:07:55.722596 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:56 crc kubenswrapper[4948]: I0220 08:07:56.722205 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:56 crc kubenswrapper[4948]: I0220 08:07:56.722232 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:56 crc kubenswrapper[4948]: E0220 08:07:56.722354 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:56 crc kubenswrapper[4948]: I0220 08:07:56.722400 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:56 crc kubenswrapper[4948]: E0220 08:07:56.722578 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:56 crc kubenswrapper[4948]: E0220 08:07:56.722678 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:57 crc kubenswrapper[4948]: E0220 08:07:57.296957 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 20 08:07:57 crc kubenswrapper[4948]: I0220 08:07:57.722519 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:57 crc kubenswrapper[4948]: E0220 08:07:57.722776 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:07:58 crc kubenswrapper[4948]: I0220 08:07:58.721864 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:07:58 crc kubenswrapper[4948]: I0220 08:07:58.722044 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:07:58 crc kubenswrapper[4948]: I0220 08:07:58.722396 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:07:58 crc kubenswrapper[4948]: E0220 08:07:58.722368 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:07:58 crc kubenswrapper[4948]: E0220 08:07:58.722679 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:07:58 crc kubenswrapper[4948]: E0220 08:07:58.722802 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:07:58 crc kubenswrapper[4948]: I0220 08:07:58.723336 4948 scope.go:117] "RemoveContainer" containerID="2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c" Feb 20 08:07:59 crc kubenswrapper[4948]: I0220 08:07:59.524424 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/1.log" Feb 20 08:07:59 crc kubenswrapper[4948]: I0220 08:07:59.524788 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7s56" event={"ID":"b96124e4-0a74-4578-9142-fd728eb9f99e","Type":"ContainerStarted","Data":"4ecf8c7c9ed34410ff4619b18254e2eb90170911300e6cd6a7284e84e8937246"} Feb 20 08:07:59 crc kubenswrapper[4948]: I0220 08:07:59.722292 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:07:59 crc kubenswrapper[4948]: E0220 08:07:59.722510 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:08:00 crc kubenswrapper[4948]: I0220 08:08:00.721705 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:00 crc kubenswrapper[4948]: I0220 08:08:00.721812 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:00 crc kubenswrapper[4948]: I0220 08:08:00.721946 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:00 crc kubenswrapper[4948]: E0220 08:08:00.722292 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:08:00 crc kubenswrapper[4948]: E0220 08:08:00.722749 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:08:00 crc kubenswrapper[4948]: E0220 08:08:00.722165 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:08:01 crc kubenswrapper[4948]: I0220 08:08:01.721707 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:01 crc kubenswrapper[4948]: E0220 08:08:01.723543 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:08:02 crc kubenswrapper[4948]: E0220 08:08:02.298519 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 20 08:08:02 crc kubenswrapper[4948]: I0220 08:08:02.722218 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:02 crc kubenswrapper[4948]: I0220 08:08:02.722258 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:02 crc kubenswrapper[4948]: I0220 08:08:02.722241 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:02 crc kubenswrapper[4948]: E0220 08:08:02.722433 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:08:02 crc kubenswrapper[4948]: E0220 08:08:02.722595 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:08:02 crc kubenswrapper[4948]: E0220 08:08:02.722731 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:08:03 crc kubenswrapper[4948]: I0220 08:08:03.722445 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:03 crc kubenswrapper[4948]: E0220 08:08:03.722673 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:08:04 crc kubenswrapper[4948]: I0220 08:08:04.722013 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:04 crc kubenswrapper[4948]: I0220 08:08:04.722076 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:04 crc kubenswrapper[4948]: I0220 08:08:04.722262 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:04 crc kubenswrapper[4948]: E0220 08:08:04.722560 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:08:04 crc kubenswrapper[4948]: E0220 08:08:04.722447 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:08:04 crc kubenswrapper[4948]: E0220 08:08:04.722246 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:08:05 crc kubenswrapper[4948]: I0220 08:08:05.721923 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:05 crc kubenswrapper[4948]: E0220 08:08:05.722213 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:08:06 crc kubenswrapper[4948]: I0220 08:08:06.721912 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:06 crc kubenswrapper[4948]: I0220 08:08:06.722031 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:06 crc kubenswrapper[4948]: I0220 08:08:06.721915 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:06 crc kubenswrapper[4948]: E0220 08:08:06.722187 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:08:06 crc kubenswrapper[4948]: E0220 08:08:06.722403 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:08:06 crc kubenswrapper[4948]: E0220 08:08:06.722558 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:08:06 crc kubenswrapper[4948]: I0220 08:08:06.723648 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:08:07 crc kubenswrapper[4948]: E0220 08:08:07.300858 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 20 08:08:07 crc kubenswrapper[4948]: I0220 08:08:07.557150 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/3.log" Feb 20 08:08:07 crc kubenswrapper[4948]: I0220 08:08:07.560418 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerStarted","Data":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} Feb 20 08:08:07 crc kubenswrapper[4948]: I0220 08:08:07.560880 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:08:07 crc kubenswrapper[4948]: I0220 08:08:07.602489 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podStartSLOduration=113.602455301 podStartE2EDuration="1m53.602455301s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:07.600146854 +0000 UTC m=+136.574641724" watchObservedRunningTime="2026-02-20 08:08:07.602455301 +0000 UTC m=+136.576950161" Feb 20 08:08:07 crc kubenswrapper[4948]: I0220 08:08:07.721941 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:07 crc kubenswrapper[4948]: E0220 08:08:07.722178 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:08:07 crc kubenswrapper[4948]: I0220 08:08:07.781854 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xd86g"] Feb 20 08:08:07 crc kubenswrapper[4948]: I0220 08:08:07.782056 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:07 crc kubenswrapper[4948]: E0220 08:08:07.782254 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:08:08 crc kubenswrapper[4948]: I0220 08:08:08.722318 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:08 crc kubenswrapper[4948]: I0220 08:08:08.722318 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:08 crc kubenswrapper[4948]: E0220 08:08:08.722567 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:08:08 crc kubenswrapper[4948]: E0220 08:08:08.722741 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:08:09 crc kubenswrapper[4948]: I0220 08:08:09.722064 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:09 crc kubenswrapper[4948]: I0220 08:08:09.722170 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:09 crc kubenswrapper[4948]: E0220 08:08:09.722263 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:08:09 crc kubenswrapper[4948]: E0220 08:08:09.722420 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:08:10 crc kubenswrapper[4948]: I0220 08:08:10.722517 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:10 crc kubenswrapper[4948]: I0220 08:08:10.722568 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:10 crc kubenswrapper[4948]: E0220 08:08:10.722742 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 08:08:10 crc kubenswrapper[4948]: E0220 08:08:10.722884 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 08:08:11 crc kubenswrapper[4948]: I0220 08:08:11.722203 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:11 crc kubenswrapper[4948]: I0220 08:08:11.722309 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:11 crc kubenswrapper[4948]: E0220 08:08:11.724027 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 08:08:11 crc kubenswrapper[4948]: E0220 08:08:11.724249 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xd86g" podUID="62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd" Feb 20 08:08:12 crc kubenswrapper[4948]: I0220 08:08:12.722394 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:12 crc kubenswrapper[4948]: I0220 08:08:12.722522 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:12 crc kubenswrapper[4948]: I0220 08:08:12.725772 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 20 08:08:12 crc kubenswrapper[4948]: I0220 08:08:12.726786 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 20 08:08:12 crc kubenswrapper[4948]: I0220 08:08:12.726797 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 20 08:08:12 crc kubenswrapper[4948]: I0220 08:08:12.727030 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 20 08:08:13 crc kubenswrapper[4948]: I0220 08:08:13.721474 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:13 crc kubenswrapper[4948]: I0220 08:08:13.721565 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:13 crc kubenswrapper[4948]: I0220 08:08:13.726231 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 20 08:08:13 crc kubenswrapper[4948]: I0220 08:08:13.726335 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.592581 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.592641 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.594178 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.600845 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.693885 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:20 crc kubenswrapper[4948]: E0220 08:08:20.694182 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:10:22.694143079 +0000 UTC m=+271.668637949 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.694545 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.694629 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.700474 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.700575 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.847811 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.861315 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 08:08:20 crc kubenswrapper[4948]: I0220 08:08:20.955405 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:21 crc kubenswrapper[4948]: W0220 08:08:21.167423 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-44f5ba7b5021740fa36055309a0fed6bca4188708200cd5ea09e952a580b5810 WatchSource:0}: Error finding container 44f5ba7b5021740fa36055309a0fed6bca4188708200cd5ea09e952a580b5810: Status 404 returned error can't find the container with id 44f5ba7b5021740fa36055309a0fed6bca4188708200cd5ea09e952a580b5810 Feb 20 08:08:21 crc kubenswrapper[4948]: W0220 08:08:21.168384 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-88ea25004576fb4b079f000bc89eaff31a4353a7cea7a707f7fddd5a0870b507 WatchSource:0}: Error finding container 88ea25004576fb4b079f000bc89eaff31a4353a7cea7a707f7fddd5a0870b507: Status 404 returned error can't find the container with id 88ea25004576fb4b079f000bc89eaff31a4353a7cea7a707f7fddd5a0870b507 Feb 20 08:08:21 crc kubenswrapper[4948]: W0220 08:08:21.206477 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-ac2e7e878b2a17066ee69c643bf29d8a5bef205795d3458c0b19eb26a5d75d5f WatchSource:0}: Error finding container ac2e7e878b2a17066ee69c643bf29d8a5bef205795d3458c0b19eb26a5d75d5f: Status 404 returned error can't find the container with id ac2e7e878b2a17066ee69c643bf29d8a5bef205795d3458c0b19eb26a5d75d5f Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.301559 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.338281 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jfst2"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.338778 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.343958 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.344003 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.344271 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.344380 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.346441 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.346705 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.346939 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.347062 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.352105 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.353709 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.354578 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.355607 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.356657 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.360872 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.362483 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.363007 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.363389 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.363943 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.364321 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.364633 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.370046 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.380732 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.381336 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.382477 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.382734 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fhfhr"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.382955 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.383246 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.383365 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.383548 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.384155 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.384418 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.384666 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.384720 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.388351 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.389245 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fthdq"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.389617 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.389757 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.393180 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.393743 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.394113 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.394334 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.394474 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.394645 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.394768 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.395276 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.395445 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.397685 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.398233 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qtlw7"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.398600 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.399744 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-xjbrz"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.399763 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.399846 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.400450 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.400579 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.400236 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.400384 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.402954 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-etcd-client\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403008 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-config\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403037 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-encryption-config\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403067 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-image-import-ca\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403089 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c7c6bf9e-0846-4391-9126-a00d68b2e627-node-pullsecrets\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403108 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403126 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb9hj\" (UniqueName: \"kubernetes.io/projected/c7c6bf9e-0846-4391-9126-a00d68b2e627-kube-api-access-hb9hj\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403149 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-audit\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403167 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-etcd-serving-ca\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403198 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-serving-cert\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.403215 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7c6bf9e-0846-4391-9126-a00d68b2e627-audit-dir\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.406100 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.406506 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.406913 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.407316 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.407839 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.408161 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.408466 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.409089 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.409902 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.410295 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.410782 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.411351 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-44l7j"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.411833 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.412185 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.412420 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.412604 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.413181 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.413398 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.413530 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414125 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414375 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414580 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414664 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414792 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-k8zcr"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414123 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414387 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414897 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.415671 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414923 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414961 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.416458 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qfgtj"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.417137 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.417243 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.416608 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.416659 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.415048 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.415086 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.415082 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.414995 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.416907 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.416951 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.430098 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.433131 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.439812 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.440312 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.441315 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-x8fgj"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.441639 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.441641 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.442011 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.459557 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.463242 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.463452 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.463541 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.463610 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.463908 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.464710 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.465114 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.465386 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.465467 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.465535 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.466664 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.467817 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.467957 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.469948 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.470074 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.470261 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.470478 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.470749 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472195 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472351 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472386 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472420 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472575 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472749 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472900 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472899 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-sc5pm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.472996 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.473099 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.473159 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.473286 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.473322 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.473516 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.474034 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.474108 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.474206 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.474359 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.474450 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.474517 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.478045 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-grnfc"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.478639 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-grnfc" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.479015 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.479355 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.481008 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.484750 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sjnxm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.485569 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.489039 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.498551 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.498628 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.499706 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.500139 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7wm8c"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.500939 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.501269 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-bcsmj"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.501883 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.505262 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.506712 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.507006 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.507371 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.507934 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.508126 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.510722 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.511198 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.511846 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.512314 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.513001 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514178 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf612779-d96d-48f1-9958-5f92ac6e3d40-serving-cert\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514201 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-client\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514219 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkhpt\" (UniqueName: \"kubernetes.io/projected/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-kube-api-access-gkhpt\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514241 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-oauth-config\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514262 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514276 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514295 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-service-ca\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514309 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq9dr\" (UniqueName: \"kubernetes.io/projected/a149fe37-c748-4120-9116-1da4b680d880-kube-api-access-nq9dr\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgm85\" (UID: \"a149fe37-c748-4120-9116-1da4b680d880\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514326 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514342 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fabbac48-e5db-4057-b9a5-69118c38c667-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514357 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514373 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz9sw\" (UniqueName: \"kubernetes.io/projected/de9b95d8-fa65-438c-bef9-2aba46044ec1-kube-api-access-mz9sw\") pod \"cluster-samples-operator-665b6dd947-7x85x\" (UID: \"de9b95d8-fa65-438c-bef9-2aba46044ec1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514388 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b1099221-b48f-4756-b133-0fed16b1e225-images\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514402 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsbfs\" (UniqueName: \"kubernetes.io/projected/21db7edd-6db3-4afa-b470-0016a9c5afb9-kube-api-access-vsbfs\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514418 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b1099221-b48f-4756-b133-0fed16b1e225-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514436 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21db7edd-6db3-4afa-b470-0016a9c5afb9-serving-cert\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514453 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02036666-815d-4282-86f0-1b4163ea7cf9-config\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514469 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514486 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-oauth-serving-cert\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514502 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkj4v\" (UniqueName: \"kubernetes.io/projected/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-kube-api-access-jkj4v\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514520 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/02036666-815d-4282-86f0-1b4163ea7cf9-machine-approver-tls\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514536 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-config\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514560 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-serving-cert\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514584 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514603 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-serving-cert\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514624 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgccc\" (UniqueName: \"kubernetes.io/projected/b1099221-b48f-4756-b133-0fed16b1e225-kube-api-access-lgccc\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514642 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-config\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514663 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514682 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mr8r\" (UniqueName: \"kubernetes.io/projected/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-kube-api-access-5mr8r\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514703 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-client-ca\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514725 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/de9b95d8-fa65-438c-bef9-2aba46044ec1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7x85x\" (UID: \"de9b95d8-fa65-438c-bef9-2aba46044ec1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514745 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-client-ca\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514761 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514776 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-config\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514791 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-config\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514806 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bt5g\" (UniqueName: \"kubernetes.io/projected/cf612779-d96d-48f1-9958-5f92ac6e3d40-kube-api-access-2bt5g\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514826 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-audit-policies\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514843 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9x9v\" (UniqueName: \"kubernetes.io/projected/2c59c4e2-97ae-494a-b33a-a542266cf233-kube-api-access-c9x9v\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514860 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514877 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94nxl\" (UniqueName: \"kubernetes.io/projected/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-kube-api-access-94nxl\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514893 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0b98dbb3-7986-475d-8028-0879d2fed2af-available-featuregates\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514911 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhx6j\" (UniqueName: \"kubernetes.io/projected/bbc3437e-8729-46f8-aee2-ad3a4679097f-kube-api-access-lhx6j\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514934 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-serving-cert\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514951 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7c6bf9e-0846-4391-9126-a00d68b2e627-audit-dir\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.514984 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbc3437e-8729-46f8-aee2-ad3a4679097f-config\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515001 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s94wm\" (UniqueName: \"kubernetes.io/projected/02036666-815d-4282-86f0-1b4163ea7cf9-kube-api-access-s94wm\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515017 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-config\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515033 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-etcd-client\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515054 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-encryption-config\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515069 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515084 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-etcd-client\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515100 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-config\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515118 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk7rx\" (UniqueName: \"kubernetes.io/projected/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-kube-api-access-lk7rx\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515133 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fabbac48-e5db-4057-b9a5-69118c38c667-config\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515147 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2c59c4e2-97ae-494a-b33a-a542266cf233-audit-dir\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515161 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-images\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515175 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515190 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-serving-cert\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515205 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b1099221-b48f-4756-b133-0fed16b1e225-proxy-tls\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515223 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-ca\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515237 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-service-ca\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515252 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bbc3437e-8729-46f8-aee2-ad3a4679097f-trusted-ca\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515267 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-config\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515446 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-encryption-config\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515495 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-trusted-ca-bundle\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515513 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a149fe37-c748-4120-9116-1da4b680d880-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgm85\" (UID: \"a149fe37-c748-4120-9116-1da4b680d880\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515531 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v824\" (UniqueName: \"kubernetes.io/projected/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-kube-api-access-2v824\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515555 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fabbac48-e5db-4057-b9a5-69118c38c667-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515576 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c070b67-c2ca-459b-a1b7-813a8833e27e-serving-cert\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515603 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-image-import-ca\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515627 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515648 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25nlb\" (UniqueName: \"kubernetes.io/projected/6c070b67-c2ca-459b-a1b7-813a8833e27e-kube-api-access-25nlb\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515666 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9k4m\" (UniqueName: \"kubernetes.io/projected/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-kube-api-access-m9k4m\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515684 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c7c6bf9e-0846-4391-9126-a00d68b2e627-node-pullsecrets\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515700 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b98dbb3-7986-475d-8028-0879d2fed2af-serving-cert\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515717 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515735 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb9hj\" (UniqueName: \"kubernetes.io/projected/c7c6bf9e-0846-4391-9126-a00d68b2e627-kube-api-access-hb9hj\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515752 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515772 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccbnv\" (UniqueName: \"kubernetes.io/projected/0b98dbb3-7986-475d-8028-0879d2fed2af-kube-api-access-ccbnv\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515791 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515807 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-proxy-tls\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515825 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xszfj\" (UniqueName: \"kubernetes.io/projected/3f4bc9ec-3eda-4d56-9696-df15c14d2e4a-kube-api-access-xszfj\") pod \"migrator-59844c95c7-cc77j\" (UID: \"3f4bc9ec-3eda-4d56-9696-df15c14d2e4a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515842 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b5rg\" (UniqueName: \"kubernetes.io/projected/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-kube-api-access-7b5rg\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515859 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-audit\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515910 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/02036666-815d-4282-86f0-1b4163ea7cf9-auth-proxy-config\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515927 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-etcd-serving-ca\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515944 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515959 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-service-ca-bundle\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.515990 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bbc3437e-8729-46f8-aee2-ad3a4679097f-serving-cert\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.516860 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.517330 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.517639 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.521453 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7c6bf9e-0846-4391-9126-a00d68b2e627-audit-dir\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.523236 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.523845 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.524059 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.524819 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-serving-cert\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.530583 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-etcd-client\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.531141 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-image-import-ca\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.531254 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c7c6bf9e-0846-4391-9126-a00d68b2e627-node-pullsecrets\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.532341 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-config\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.536297 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-etcd-serving-ca\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.536738 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-audit\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.538525 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7c6bf9e-0846-4391-9126-a00d68b2e627-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.538579 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rtnmx"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.542458 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.542473 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.544872 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c7c6bf9e-0846-4391-9126-a00d68b2e627-encryption-config\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.546737 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sjv8t"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.547002 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.548491 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jfst2"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.548605 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.549924 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.560034 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fhfhr"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.562751 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.564231 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.565268 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.566753 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.569012 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fthdq"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.570506 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.572426 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-xjbrz"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.574720 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.576321 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.577827 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-9ls2l"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.578749 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.579922 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-bcsmj"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.580842 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-44l7j"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.582292 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-bp2vx"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.583557 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.584036 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.584045 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-bp2vx" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.585369 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-grnfc"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.586996 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-k8zcr"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.589660 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.591636 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qtlw7"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.593213 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.594485 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.595757 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.600842 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7wm8c"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.602048 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.603883 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qfgtj"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.605310 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.606678 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.608187 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-x8fgj"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.609349 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.610925 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.612286 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.613949 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sjnxm"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.615374 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-trh7g"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616733 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bbc3437e-8729-46f8-aee2-ad3a4679097f-serving-cert\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616775 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616801 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e9f46a23-4535-4d64-ae76-065435f8f762-metrics-tls\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616824 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616843 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-service-ca-bundle\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616871 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616892 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf612779-d96d-48f1-9958-5f92ac6e3d40-serving-cert\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616915 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-client\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616935 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkhpt\" (UniqueName: \"kubernetes.io/projected/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-kube-api-access-gkhpt\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616953 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-oauth-config\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.616982 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617002 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-srv-cert\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617036 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2l9z\" (UniqueName: \"kubernetes.io/projected/df73228e-df1f-4e9c-aafe-493bb743f98a-kube-api-access-q2l9z\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617062 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fabbac48-e5db-4057-b9a5-69118c38c667-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617087 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-service-ca\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617117 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq9dr\" (UniqueName: \"kubernetes.io/projected/a149fe37-c748-4120-9116-1da4b680d880-kube-api-access-nq9dr\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgm85\" (UID: \"a149fe37-c748-4120-9116-1da4b680d880\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617143 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617167 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b1099221-b48f-4756-b133-0fed16b1e225-images\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617184 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617202 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mz9sw\" (UniqueName: \"kubernetes.io/projected/de9b95d8-fa65-438c-bef9-2aba46044ec1-kube-api-access-mz9sw\") pod \"cluster-samples-operator-665b6dd947-7x85x\" (UID: \"de9b95d8-fa65-438c-bef9-2aba46044ec1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617228 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsbfs\" (UniqueName: \"kubernetes.io/projected/21db7edd-6db3-4afa-b470-0016a9c5afb9-kube-api-access-vsbfs\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617247 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617273 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02036666-815d-4282-86f0-1b4163ea7cf9-config\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617359 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b1099221-b48f-4756-b133-0fed16b1e225-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617810 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21db7edd-6db3-4afa-b470-0016a9c5afb9-serving-cert\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617864 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmpf7\" (UniqueName: \"kubernetes.io/projected/e9f46a23-4535-4d64-ae76-065435f8f762-kube-api-access-lmpf7\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617904 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrvhw\" (UniqueName: \"kubernetes.io/projected/bc13d9fe-fd58-4d3b-9278-4f3e59da5976-kube-api-access-lrvhw\") pod \"downloads-7954f5f757-grnfc\" (UID: \"bc13d9fe-fd58-4d3b-9278-4f3e59da5976\") " pod="openshift-console/downloads-7954f5f757-grnfc" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617934 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.617960 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-policies\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618002 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/02036666-815d-4282-86f0-1b4163ea7cf9-machine-approver-tls\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618049 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-oauth-serving-cert\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618078 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkj4v\" (UniqueName: \"kubernetes.io/projected/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-kube-api-access-jkj4v\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618103 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618131 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618152 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-config\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618173 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-serving-cert\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618192 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618208 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-serving-cert\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618226 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgccc\" (UniqueName: \"kubernetes.io/projected/b1099221-b48f-4756-b133-0fed16b1e225-kube-api-access-lgccc\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618244 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-config\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618262 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618282 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-client-ca\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618298 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mr8r\" (UniqueName: \"kubernetes.io/projected/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-kube-api-access-5mr8r\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618317 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-client-ca\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618338 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hqgh\" (UniqueName: \"kubernetes.io/projected/6f1e0caa-03ee-4025-9f2a-6941090c178b-kube-api-access-6hqgh\") pod \"dns-operator-744455d44c-x8fgj\" (UID: \"6f1e0caa-03ee-4025-9f2a-6941090c178b\") " pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618364 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/de9b95d8-fa65-438c-bef9-2aba46044ec1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7x85x\" (UID: \"de9b95d8-fa65-438c-bef9-2aba46044ec1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618383 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618400 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618418 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618438 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-config\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618448 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02036666-815d-4282-86f0-1b4163ea7cf9-config\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618457 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618500 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-config\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618520 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94nxl\" (UniqueName: \"kubernetes.io/projected/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-kube-api-access-94nxl\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618538 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bt5g\" (UniqueName: \"kubernetes.io/projected/cf612779-d96d-48f1-9958-5f92ac6e3d40-kube-api-access-2bt5g\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618557 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-audit-policies\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618564 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-gd69n"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618577 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9x9v\" (UniqueName: \"kubernetes.io/projected/2c59c4e2-97ae-494a-b33a-a542266cf233-kube-api-access-c9x9v\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618598 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618821 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618568 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618945 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.618996 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0b98dbb3-7986-475d-8028-0879d2fed2af-available-featuregates\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619021 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9f46a23-4535-4d64-ae76-065435f8f762-trusted-ca\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619042 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbc3437e-8729-46f8-aee2-ad3a4679097f-config\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619061 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhx6j\" (UniqueName: \"kubernetes.io/projected/bbc3437e-8729-46f8-aee2-ad3a4679097f-kube-api-access-lhx6j\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619079 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619098 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n92xj\" (UniqueName: \"kubernetes.io/projected/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-kube-api-access-n92xj\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619119 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s94wm\" (UniqueName: \"kubernetes.io/projected/02036666-815d-4282-86f0-1b4163ea7cf9-kube-api-access-s94wm\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619134 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-config\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619151 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-etcd-client\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619168 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-encryption-config\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619188 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619198 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b1099221-b48f-4756-b133-0fed16b1e225-images\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619214 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-dir\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619247 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk7rx\" (UniqueName: \"kubernetes.io/projected/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-kube-api-access-lk7rx\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619273 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fabbac48-e5db-4057-b9a5-69118c38c667-config\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619295 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2c59c4e2-97ae-494a-b33a-a542266cf233-audit-dir\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619311 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619307 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-service-ca\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619332 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b1099221-b48f-4756-b133-0fed16b1e225-proxy-tls\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619517 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-images\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619538 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619556 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-serving-cert\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619582 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6f1e0caa-03ee-4025-9f2a-6941090c178b-metrics-tls\") pod \"dns-operator-744455d44c-x8fgj\" (UID: \"6f1e0caa-03ee-4025-9f2a-6941090c178b\") " pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619598 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619619 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-ca\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619636 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-service-ca\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619652 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bbc3437e-8729-46f8-aee2-ad3a4679097f-trusted-ca\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619671 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a149fe37-c748-4120-9116-1da4b680d880-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgm85\" (UID: \"a149fe37-c748-4120-9116-1da4b680d880\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619690 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-config\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619710 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-trusted-ca-bundle\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619733 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619754 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v824\" (UniqueName: \"kubernetes.io/projected/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-kube-api-access-2v824\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619773 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fabbac48-e5db-4057-b9a5-69118c38c667-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619794 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c070b67-c2ca-459b-a1b7-813a8833e27e-serving-cert\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619827 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25nlb\" (UniqueName: \"kubernetes.io/projected/6c070b67-c2ca-459b-a1b7-813a8833e27e-kube-api-access-25nlb\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619825 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-client-ca\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619846 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9k4m\" (UniqueName: \"kubernetes.io/projected/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-kube-api-access-m9k4m\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619868 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b98dbb3-7986-475d-8028-0879d2fed2af-serving-cert\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619893 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619915 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccbnv\" (UniqueName: \"kubernetes.io/projected/0b98dbb3-7986-475d-8028-0879d2fed2af-kube-api-access-ccbnv\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619935 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619952 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.620011 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b5rg\" (UniqueName: \"kubernetes.io/projected/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-kube-api-access-7b5rg\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.620032 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.620054 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-proxy-tls\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.619692 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-config\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.621065 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.621448 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-oauth-config\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.621861 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-config\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.621913 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xszfj\" (UniqueName: \"kubernetes.io/projected/3f4bc9ec-3eda-4d56-9696-df15c14d2e4a-kube-api-access-xszfj\") pod \"migrator-59844c95c7-cc77j\" (UID: \"3f4bc9ec-3eda-4d56-9696-df15c14d2e4a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.621942 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9f46a23-4535-4d64-ae76-065435f8f762-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.621969 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/02036666-815d-4282-86f0-1b4163ea7cf9-auth-proxy-config\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.622025 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-audit-policies\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.622483 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/02036666-815d-4282-86f0-1b4163ea7cf9-auth-proxy-config\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.622492 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-client\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.622768 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.622818 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.622956 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.623271 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-serving-cert\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.623673 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b1099221-b48f-4756-b133-0fed16b1e225-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.623744 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-ca\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.624214 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"6d33a120c6d6404bfdafb94ac3396204dc5d916bd0861e293668c90e52a5a928"} Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.624269 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.624295 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rtnmx"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.624310 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ac2e7e878b2a17066ee69c643bf29d8a5bef205795d3458c0b19eb26a5d75d5f"} Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.624324 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"448b4e1f87f68cbd93b97818fd8a7931f478e3d00887e8cf4aadad7e8d891215"} Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.624337 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"88ea25004576fb4b079f000bc89eaff31a4353a7cea7a707f7fddd5a0870b507"} Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.624622 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-config\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.624837 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bbc3437e-8729-46f8-aee2-ad3a4679097f-serving-cert\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.625016 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.625122 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c59c4e2-97ae-494a-b33a-a542266cf233-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.625227 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a16ed2bfb92b81fce0adc99ac98001bff53b82cfd8c212217dee3d7618503946"} Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.625289 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"44f5ba7b5021740fa36055309a0fed6bca4188708200cd5ea09e952a580b5810"} Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.625331 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.625625 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-config\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.627347 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-config\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.627946 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-proxy-tls\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.628218 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/02036666-815d-4282-86f0-1b4163ea7cf9-machine-approver-tls\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.628302 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-oauth-serving-cert\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.628582 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fabbac48-e5db-4057-b9a5-69118c38c667-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.629102 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b1099221-b48f-4756-b133-0fed16b1e225-proxy-tls\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.629650 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-images\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.630388 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-encryption-config\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.630559 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.630715 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-etcd-client\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.630930 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b98dbb3-7986-475d-8028-0879d2fed2af-serving-cert\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.630998 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fabbac48-e5db-4057-b9a5-69118c38c667-config\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.631030 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2c59c4e2-97ae-494a-b33a-a542266cf233-audit-dir\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.631145 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0b98dbb3-7986-475d-8028-0879d2fed2af-available-featuregates\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.631271 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf612779-d96d-48f1-9958-5f92ac6e3d40-serving-cert\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.631301 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.631498 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-service-ca-bundle\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.631512 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-trusted-ca-bundle\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.631539 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-9ls2l"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.632485 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.632491 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.632795 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cf612779-d96d-48f1-9958-5f92ac6e3d40-etcd-service-ca\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.633038 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.633156 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-client-ca\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.633172 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.634077 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bbc3437e-8729-46f8-aee2-ad3a4679097f-trusted-ca\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.634075 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c59c4e2-97ae-494a-b33a-a542266cf233-serving-cert\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.634268 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.634312 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-config\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.634678 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-serving-cert\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.635204 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/de9b95d8-fa65-438c-bef9-2aba46044ec1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7x85x\" (UID: \"de9b95d8-fa65-438c-bef9-2aba46044ec1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.635294 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-bp2vx"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.642272 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sjv8t"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.642314 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.642327 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.643930 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21db7edd-6db3-4afa-b470-0016a9c5afb9-serving-cert\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.645469 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.646138 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbc3437e-8729-46f8-aee2-ad3a4679097f-config\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.648416 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.649091 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c070b67-c2ca-459b-a1b7-813a8833e27e-serving-cert\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.652700 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-trh7g"] Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.662277 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.681728 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.687692 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a149fe37-c748-4120-9116-1da4b680d880-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgm85\" (UID: \"a149fe37-c748-4120-9116-1da4b680d880\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.702935 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.722026 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.722863 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-dir\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723007 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723103 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-dir\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723107 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6f1e0caa-03ee-4025-9f2a-6941090c178b-metrics-tls\") pod \"dns-operator-744455d44c-x8fgj\" (UID: \"6f1e0caa-03ee-4025-9f2a-6941090c178b\") " pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723173 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723265 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723307 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723350 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9f46a23-4535-4d64-ae76-065435f8f762-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723381 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723403 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e9f46a23-4535-4d64-ae76-065435f8f762-metrics-tls\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723440 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-srv-cert\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723466 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2l9z\" (UniqueName: \"kubernetes.io/projected/df73228e-df1f-4e9c-aafe-493bb743f98a-kube-api-access-q2l9z\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723532 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723566 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmpf7\" (UniqueName: \"kubernetes.io/projected/e9f46a23-4535-4d64-ae76-065435f8f762-kube-api-access-lmpf7\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723589 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrvhw\" (UniqueName: \"kubernetes.io/projected/bc13d9fe-fd58-4d3b-9278-4f3e59da5976-kube-api-access-lrvhw\") pod \"downloads-7954f5f757-grnfc\" (UID: \"bc13d9fe-fd58-4d3b-9278-4f3e59da5976\") " pod="openshift-console/downloads-7954f5f757-grnfc" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723612 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-policies\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723644 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723689 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723746 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hqgh\" (UniqueName: \"kubernetes.io/projected/6f1e0caa-03ee-4025-9f2a-6941090c178b-kube-api-access-6hqgh\") pod \"dns-operator-744455d44c-x8fgj\" (UID: \"6f1e0caa-03ee-4025-9f2a-6941090c178b\") " pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723778 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723799 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723822 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723882 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9f46a23-4535-4d64-ae76-065435f8f762-trusted-ca\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723911 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.723932 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n92xj\" (UniqueName: \"kubernetes.io/projected/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-kube-api-access-n92xj\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.740533 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.761640 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.769069 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6f1e0caa-03ee-4025-9f2a-6941090c178b-metrics-tls\") pod \"dns-operator-744455d44c-x8fgj\" (UID: \"6f1e0caa-03ee-4025-9f2a-6941090c178b\") " pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.781384 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.821546 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.841685 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.860930 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.881668 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.902803 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.921535 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.943002 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.961615 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 20 08:08:21 crc kubenswrapper[4948]: I0220 08:08:21.981572 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.001037 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.022645 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.041412 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.050256 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.062769 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.081950 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.090416 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.100632 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.109139 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.149769 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.155600 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.160713 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.161639 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.166990 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.169275 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.181258 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.187255 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.204899 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.219530 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.241624 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.243266 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.246352 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.261466 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.264994 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-policies\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.282108 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.301527 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.304546 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.321304 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.325359 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.342052 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.361767 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.382600 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.402357 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.421749 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.442133 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.461843 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.481073 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.487703 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.501485 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.520240 4948 request.go:700] Waited for 1.011864135s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dpackageserver-service-cert&limit=500&resourceVersion=0 Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.522806 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.541258 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.569195 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.576180 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9f46a23-4535-4d64-ae76-065435f8f762-trusted-ca\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.581827 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.601832 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.610083 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e9f46a23-4535-4d64-ae76-065435f8f762-metrics-tls\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.622708 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.641420 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.662161 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.682058 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.701264 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.708891 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-srv-cert\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.725130 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.743097 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.761532 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.781351 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.802904 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.821996 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.842073 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.861228 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.881740 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.902601 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.942075 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.952597 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb9hj\" (UniqueName: \"kubernetes.io/projected/c7c6bf9e-0846-4391-9126-a00d68b2e627-kube-api-access-hb9hj\") pod \"apiserver-76f77b778f-jfst2\" (UID: \"c7c6bf9e-0846-4391-9126-a00d68b2e627\") " pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.961496 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 20 08:08:22 crc kubenswrapper[4948]: I0220 08:08:22.982094 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.002050 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.022353 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.041914 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.061591 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.093046 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.102204 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.122838 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.161656 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.180067 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.183586 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.203090 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.223579 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.245624 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.261539 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.281292 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.339059 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fabbac48-e5db-4057-b9a5-69118c38c667-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vtkbz\" (UID: \"fabbac48-e5db-4057-b9a5-69118c38c667\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.354105 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkhpt\" (UniqueName: \"kubernetes.io/projected/1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72-kube-api-access-gkhpt\") pod \"machine-config-controller-84d6567774-8rrhs\" (UID: \"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.375423 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq9dr\" (UniqueName: \"kubernetes.io/projected/a149fe37-c748-4120-9116-1da4b680d880-kube-api-access-nq9dr\") pod \"control-plane-machine-set-operator-78cbb6b69f-bgm85\" (UID: \"a149fe37-c748-4120-9116-1da4b680d880\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.381724 4948 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.391807 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.394022 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.402645 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.422061 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.463723 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jfst2"] Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.466369 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mr8r\" (UniqueName: \"kubernetes.io/projected/5bc3d0cd-eacd-4b25-9acd-853e49db7b47-kube-api-access-5mr8r\") pod \"machine-api-operator-5694c8668f-fthdq\" (UID: \"5bc3d0cd-eacd-4b25-9acd-853e49db7b47\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.493057 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94nxl\" (UniqueName: \"kubernetes.io/projected/e029f4ab-6a61-4fcf-9997-b51c1cbe5675-kube-api-access-94nxl\") pod \"kube-storage-version-migrator-operator-b67b599dd-nzj8d\" (UID: \"e029f4ab-6a61-4fcf-9997-b51c1cbe5675\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.509245 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bt5g\" (UniqueName: \"kubernetes.io/projected/cf612779-d96d-48f1-9958-5f92ac6e3d40-kube-api-access-2bt5g\") pod \"etcd-operator-b45778765-qfgtj\" (UID: \"cf612779-d96d-48f1-9958-5f92ac6e3d40\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.509684 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.514067 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.522782 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mz9sw\" (UniqueName: \"kubernetes.io/projected/de9b95d8-fa65-438c-bef9-2aba46044ec1-kube-api-access-mz9sw\") pod \"cluster-samples-operator-665b6dd947-7x85x\" (UID: \"de9b95d8-fa65-438c-bef9-2aba46044ec1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.529516 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.539620 4948 request.go:700] Waited for 1.915562757s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/serviceaccounts/machine-config-operator/token Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.541153 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v824\" (UniqueName: \"kubernetes.io/projected/01d95eeb-e421-4fe2-a24a-44c3014f1f6e-kube-api-access-2v824\") pod \"cluster-image-registry-operator-dc59b4c8b-thcgl\" (UID: \"01d95eeb-e421-4fe2-a24a-44c3014f1f6e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.558237 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgccc\" (UniqueName: \"kubernetes.io/projected/b1099221-b48f-4756-b133-0fed16b1e225-kube-api-access-lgccc\") pod \"machine-config-operator-74547568cd-lrdf2\" (UID: \"b1099221-b48f-4756-b133-0fed16b1e225\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.561513 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.570430 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.578137 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.581487 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.616433 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.648934 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b5rg\" (UniqueName: \"kubernetes.io/projected/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-kube-api-access-7b5rg\") pod \"console-f9d7485db-k8zcr\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.655010 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz"] Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.662369 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.667539 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" event={"ID":"c7c6bf9e-0846-4391-9126-a00d68b2e627","Type":"ContainerStarted","Data":"810c77e885fc3d0ade548eab4f913ad0847c543d802265ae386fa74cc753a8d0"} Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.668083 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkj4v\" (UniqueName: \"kubernetes.io/projected/05fe2d3b-0a53-4bbe-b0aa-be07acb159bb-kube-api-access-jkj4v\") pod \"authentication-operator-69f744f599-qtlw7\" (UID: \"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.678911 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.695637 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk7rx\" (UniqueName: \"kubernetes.io/projected/0b81d26a-f92a-4602-89ef-5dd9fb24a32f-kube-api-access-lk7rx\") pod \"openshift-controller-manager-operator-756b6f6bc6-klsxg\" (UID: \"0b81d26a-f92a-4602-89ef-5dd9fb24a32f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.700429 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9x9v\" (UniqueName: \"kubernetes.io/projected/2c59c4e2-97ae-494a-b33a-a542266cf233-kube-api-access-c9x9v\") pod \"apiserver-7bbb656c7d-gf42w\" (UID: \"2c59c4e2-97ae-494a-b33a-a542266cf233\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:23 crc kubenswrapper[4948]: W0220 08:08:23.714071 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfabbac48_e5db_4057_b9a5_69118c38c667.slice/crio-0078d6f35d4b09153ac5f4519a90739d3b7c1a3332b872561f2b13e7ca2d48f9 WatchSource:0}: Error finding container 0078d6f35d4b09153ac5f4519a90739d3b7c1a3332b872561f2b13e7ca2d48f9: Status 404 returned error can't find the container with id 0078d6f35d4b09153ac5f4519a90739d3b7c1a3332b872561f2b13e7ca2d48f9 Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.716807 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccbnv\" (UniqueName: \"kubernetes.io/projected/0b98dbb3-7986-475d-8028-0879d2fed2af-kube-api-access-ccbnv\") pod \"openshift-config-operator-7777fb866f-44l7j\" (UID: \"0b98dbb3-7986-475d-8028-0879d2fed2af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.734956 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xszfj\" (UniqueName: \"kubernetes.io/projected/3f4bc9ec-3eda-4d56-9696-df15c14d2e4a-kube-api-access-xszfj\") pod \"migrator-59844c95c7-cc77j\" (UID: \"3f4bc9ec-3eda-4d56-9696-df15c14d2e4a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.760552 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhx6j\" (UniqueName: \"kubernetes.io/projected/bbc3437e-8729-46f8-aee2-ad3a4679097f-kube-api-access-lhx6j\") pod \"console-operator-58897d9998-xjbrz\" (UID: \"bbc3437e-8729-46f8-aee2-ad3a4679097f\") " pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.770093 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.776837 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.779602 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25nlb\" (UniqueName: \"kubernetes.io/projected/6c070b67-c2ca-459b-a1b7-813a8833e27e-kube-api-access-25nlb\") pod \"route-controller-manager-6576b87f9c-qljmk\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.784926 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.796599 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.800705 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.802615 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s94wm\" (UniqueName: \"kubernetes.io/projected/02036666-815d-4282-86f0-1b4163ea7cf9-kube-api-access-s94wm\") pod \"machine-approver-56656f9798-v4kfs\" (UID: \"02036666-815d-4282-86f0-1b4163ea7cf9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.815438 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.820472 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.828505 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9k4m\" (UniqueName: \"kubernetes.io/projected/c38c7edf-ca04-4ea4-b15f-d83dea3f545e-kube-api-access-m9k4m\") pod \"openshift-apiserver-operator-796bbdcf4f-fl47j\" (UID: \"c38c7edf-ca04-4ea4-b15f-d83dea3f545e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.833862 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.836396 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsbfs\" (UniqueName: \"kubernetes.io/projected/21db7edd-6db3-4afa-b470-0016a9c5afb9-kube-api-access-vsbfs\") pod \"controller-manager-879f6c89f-fhfhr\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.874661 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9f46a23-4535-4d64-ae76-065435f8f762-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.877673 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2l9z\" (UniqueName: \"kubernetes.io/projected/df73228e-df1f-4e9c-aafe-493bb743f98a-kube-api-access-q2l9z\") pod \"oauth-openshift-558db77b4-sjnxm\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.892157 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.906138 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrvhw\" (UniqueName: \"kubernetes.io/projected/bc13d9fe-fd58-4d3b-9278-4f3e59da5976-kube-api-access-lrvhw\") pod \"downloads-7954f5f757-grnfc\" (UID: \"bc13d9fe-fd58-4d3b-9278-4f3e59da5976\") " pod="openshift-console/downloads-7954f5f757-grnfc" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.926788 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmpf7\" (UniqueName: \"kubernetes.io/projected/e9f46a23-4535-4d64-ae76-065435f8f762-kube-api-access-lmpf7\") pod \"ingress-operator-5b745b69d9-6jlbw\" (UID: \"e9f46a23-4535-4d64-ae76-065435f8f762\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.930350 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.937600 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl"] Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.941868 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hqgh\" (UniqueName: \"kubernetes.io/projected/6f1e0caa-03ee-4025-9f2a-6941090c178b-kube-api-access-6hqgh\") pod \"dns-operator-744455d44c-x8fgj\" (UID: \"6f1e0caa-03ee-4025-9f2a-6941090c178b\") " pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.964851 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n92xj\" (UniqueName: \"kubernetes.io/projected/c7b67dd6-4025-45ad-98f8-2e7f8ab4b573-kube-api-access-n92xj\") pod \"olm-operator-6b444d44fb-6zdkm\" (UID: \"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.975878 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" Feb 20 08:08:23 crc kubenswrapper[4948]: I0220 08:08:23.992675 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fthdq"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.008504 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.010688 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.011195 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qfgtj"] Feb 20 08:08:24 crc kubenswrapper[4948]: W0220 08:08:24.023658 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01d95eeb_e421_4fe2_a24a_44c3014f1f6e.slice/crio-fd9df30c5d63b531a8098b8313e0aead45e05e65f20b2d3a9242e2f927d8db1f WatchSource:0}: Error finding container fd9df30c5d63b531a8098b8313e0aead45e05e65f20b2d3a9242e2f927d8db1f: Status 404 returned error can't find the container with id fd9df30c5d63b531a8098b8313e0aead45e05e65f20b2d3a9242e2f927d8db1f Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.055623 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.071774 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088111 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/308389f1-6962-45be-87b0-1b061caa2c50-signing-key\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088450 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmt4n\" (UniqueName: \"kubernetes.io/projected/912a3267-581f-47b1-873c-b0e4d8dd3768-kube-api-access-zmt4n\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088494 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb7c802e-5175-4b0b-8b70-91efb1c83fa1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-46sg8\" (UID: \"fb7c802e-5175-4b0b-8b70-91efb1c83fa1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088536 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99fafd36-91d7-45b8-a321-59307d79e5ce-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088598 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55q6w\" (UniqueName: \"kubernetes.io/projected/95a270e5-7a08-4238-a65d-96986f2d2c39-kube-api-access-55q6w\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088617 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npgsj\" (UniqueName: \"kubernetes.io/projected/000483d2-8b2d-4403-9652-d4d86d55f7b5-kube-api-access-npgsj\") pod \"multus-admission-controller-857f4d67dd-bcsmj\" (UID: \"000483d2-8b2d-4403-9652-d4d86d55f7b5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088672 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-metrics-certs\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088689 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj9kq\" (UniqueName: \"kubernetes.io/projected/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-kube-api-access-fj9kq\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088705 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99fafd36-91d7-45b8-a321-59307d79e5ce-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088724 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40656b16-31ff-4065-b87c-5678308e9fb4-config\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088744 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/57c455f7-aa9c-405c-bee3-89726b84f5db-installation-pull-secrets\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088778 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99fafd36-91d7-45b8-a321-59307d79e5ce-config\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088806 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a102d1fa-69a5-4826-ac40-e07ab4087558-apiservice-cert\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088825 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-certificates\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088860 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088874 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/912a3267-581f-47b1-873c-b0e4d8dd3768-srv-cert\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088915 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsw92\" (UniqueName: \"kubernetes.io/projected/fb7c802e-5175-4b0b-8b70-91efb1c83fa1-kube-api-access-nsw92\") pod \"package-server-manager-789f6589d5-46sg8\" (UID: \"fb7c802e-5175-4b0b-8b70-91efb1c83fa1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.088935 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vg4v\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-kube-api-access-2vg4v\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089009 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-stats-auth\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089044 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089095 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089114 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089131 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-default-certificate\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089164 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjwcj\" (UniqueName: \"kubernetes.io/projected/308389f1-6962-45be-87b0-1b061caa2c50-kube-api-access-vjwcj\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089179 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089195 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089211 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f77016a6-cd35-49df-b6b5-65b4858b41c9-config-volume\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089229 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/000483d2-8b2d-4403-9652-d4d86d55f7b5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-bcsmj\" (UID: \"000483d2-8b2d-4403-9652-d4d86d55f7b5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089253 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-tls\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089287 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svs7x\" (UniqueName: \"kubernetes.io/projected/a102d1fa-69a5-4826-ac40-e07ab4087558-kube-api-access-svs7x\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089303 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/912a3267-581f-47b1-873c-b0e4d8dd3768-profile-collector-cert\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089321 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ql4r\" (UniqueName: \"kubernetes.io/projected/f77016a6-cd35-49df-b6b5-65b4858b41c9-kube-api-access-2ql4r\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089358 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-trusted-ca\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089406 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-bound-sa-token\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089423 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a102d1fa-69a5-4826-ac40-e07ab4087558-tmpfs\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089457 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/57c455f7-aa9c-405c-bee3-89726b84f5db-ca-trust-extracted\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089473 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/308389f1-6962-45be-87b0-1b061caa2c50-signing-cabundle\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089500 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a102d1fa-69a5-4826-ac40-e07ab4087558-webhook-cert\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089524 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-service-ca-bundle\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089547 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng7x4\" (UniqueName: \"kubernetes.io/projected/40656b16-31ff-4065-b87c-5678308e9fb4-kube-api-access-ng7x4\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089666 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40656b16-31ff-4065-b87c-5678308e9fb4-serving-cert\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.089719 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f77016a6-cd35-49df-b6b5-65b4858b41c9-secret-volume\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.091957 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:24.591937614 +0000 UTC m=+153.566432434 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.136920 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.168285 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-grnfc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.169477 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.190412 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.190659 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:24.690627935 +0000 UTC m=+153.665122755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.190992 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a102d1fa-69a5-4826-ac40-e07ab4087558-webhook-cert\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191021 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-service-ca-bundle\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191043 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng7x4\" (UniqueName: \"kubernetes.io/projected/40656b16-31ff-4065-b87c-5678308e9fb4-kube-api-access-ng7x4\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191067 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40656b16-31ff-4065-b87c-5678308e9fb4-serving-cert\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191111 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f77016a6-cd35-49df-b6b5-65b4858b41c9-secret-volume\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191133 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/308389f1-6962-45be-87b0-1b061caa2c50-signing-key\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191169 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pd7s\" (UniqueName: \"kubernetes.io/projected/9dbe3199-893b-4d67-a447-fa3223c321e3-kube-api-access-9pd7s\") pod \"ingress-canary-bp2vx\" (UID: \"9dbe3199-893b-4d67-a447-fa3223c321e3\") " pod="openshift-ingress-canary/ingress-canary-bp2vx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191202 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmt4n\" (UniqueName: \"kubernetes.io/projected/912a3267-581f-47b1-873c-b0e4d8dd3768-kube-api-access-zmt4n\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191243 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb7c802e-5175-4b0b-8b70-91efb1c83fa1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-46sg8\" (UID: \"fb7c802e-5175-4b0b-8b70-91efb1c83fa1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191267 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-node-bootstrap-token\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191297 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99fafd36-91d7-45b8-a321-59307d79e5ce-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191340 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55q6w\" (UniqueName: \"kubernetes.io/projected/95a270e5-7a08-4238-a65d-96986f2d2c39-kube-api-access-55q6w\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191371 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npgsj\" (UniqueName: \"kubernetes.io/projected/000483d2-8b2d-4403-9652-d4d86d55f7b5-kube-api-access-npgsj\") pod \"multus-admission-controller-857f4d67dd-bcsmj\" (UID: \"000483d2-8b2d-4403-9652-d4d86d55f7b5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191395 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv78n\" (UniqueName: \"kubernetes.io/projected/06e36dcf-9973-4369-8dc0-735cc42a68aa-kube-api-access-qv78n\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191428 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-csi-data-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191481 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-metrics-certs\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191501 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06e36dcf-9973-4369-8dc0-735cc42a68aa-config-volume\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191522 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj9kq\" (UniqueName: \"kubernetes.io/projected/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-kube-api-access-fj9kq\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191545 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99fafd36-91d7-45b8-a321-59307d79e5ce-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191563 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40656b16-31ff-4065-b87c-5678308e9fb4-config\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191595 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/57c455f7-aa9c-405c-bee3-89726b84f5db-installation-pull-secrets\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191615 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99fafd36-91d7-45b8-a321-59307d79e5ce-config\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191633 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/06e36dcf-9973-4369-8dc0-735cc42a68aa-metrics-tls\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191670 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a102d1fa-69a5-4826-ac40-e07ab4087558-apiservice-cert\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191711 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-certificates\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191742 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191761 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/912a3267-581f-47b1-873c-b0e4d8dd3768-srv-cert\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191779 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsw92\" (UniqueName: \"kubernetes.io/projected/fb7c802e-5175-4b0b-8b70-91efb1c83fa1-kube-api-access-nsw92\") pod \"package-server-manager-789f6589d5-46sg8\" (UID: \"fb7c802e-5175-4b0b-8b70-91efb1c83fa1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191823 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vg4v\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-kube-api-access-2vg4v\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191905 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-stats-auth\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191934 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.191953 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-certs\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192027 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192047 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192066 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-default-certificate\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192100 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjwcj\" (UniqueName: \"kubernetes.io/projected/308389f1-6962-45be-87b0-1b061caa2c50-kube-api-access-vjwcj\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192118 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-socket-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192137 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192155 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtrdt\" (UniqueName: \"kubernetes.io/projected/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-kube-api-access-qtrdt\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192210 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192841 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f77016a6-cd35-49df-b6b5-65b4858b41c9-config-volume\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192872 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/000483d2-8b2d-4403-9652-d4d86d55f7b5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-bcsmj\" (UID: \"000483d2-8b2d-4403-9652-d4d86d55f7b5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192914 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-tls\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192962 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svs7x\" (UniqueName: \"kubernetes.io/projected/a102d1fa-69a5-4826-ac40-e07ab4087558-kube-api-access-svs7x\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.192992 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/912a3267-581f-47b1-873c-b0e4d8dd3768-profile-collector-cert\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193021 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-plugins-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193049 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-mountpoint-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193079 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ql4r\" (UniqueName: \"kubernetes.io/projected/f77016a6-cd35-49df-b6b5-65b4858b41c9-kube-api-access-2ql4r\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193097 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9dbe3199-893b-4d67-a447-fa3223c321e3-cert\") pod \"ingress-canary-bp2vx\" (UID: \"9dbe3199-893b-4d67-a447-fa3223c321e3\") " pod="openshift-ingress-canary/ingress-canary-bp2vx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193115 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-registration-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193132 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5jbw\" (UniqueName: \"kubernetes.io/projected/13b4eba0-42e7-448a-998c-d7aaeffc9218-kube-api-access-v5jbw\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193154 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-trusted-ca\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193212 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-bound-sa-token\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193228 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a102d1fa-69a5-4826-ac40-e07ab4087558-tmpfs\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193256 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/57c455f7-aa9c-405c-bee3-89726b84f5db-ca-trust-extracted\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.193276 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/308389f1-6962-45be-87b0-1b061caa2c50-signing-cabundle\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.194544 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-service-ca-bundle\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.195638 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-certificates\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.196146 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:24.69613101 +0000 UTC m=+153.670625830 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.197790 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-trusted-ca\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.198485 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a102d1fa-69a5-4826-ac40-e07ab4087558-tmpfs\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.198723 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/57c455f7-aa9c-405c-bee3-89726b84f5db-ca-trust-extracted\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.199284 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/308389f1-6962-45be-87b0-1b061caa2c50-signing-cabundle\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.201369 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40656b16-31ff-4065-b87c-5678308e9fb4-config\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.201818 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40656b16-31ff-4065-b87c-5678308e9fb4-serving-cert\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.202171 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.203825 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.203941 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.204248 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a102d1fa-69a5-4826-ac40-e07ab4087558-apiservice-cert\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.205388 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-stats-auth\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.206505 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f77016a6-cd35-49df-b6b5-65b4858b41c9-config-volume\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.208446 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/57c455f7-aa9c-405c-bee3-89726b84f5db-installation-pull-secrets\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.210617 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/308389f1-6962-45be-87b0-1b061caa2c50-signing-key\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.211635 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/912a3267-581f-47b1-873c-b0e4d8dd3768-srv-cert\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.212007 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb7c802e-5175-4b0b-8b70-91efb1c83fa1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-46sg8\" (UID: \"fb7c802e-5175-4b0b-8b70-91efb1c83fa1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.212489 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a102d1fa-69a5-4826-ac40-e07ab4087558-webhook-cert\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.212597 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-default-certificate\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.213186 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/912a3267-581f-47b1-873c-b0e4d8dd3768-profile-collector-cert\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.213480 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99fafd36-91d7-45b8-a321-59307d79e5ce-config\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.214258 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-metrics-certs\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.214346 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99fafd36-91d7-45b8-a321-59307d79e5ce-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.219221 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.227768 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-tls\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.227990 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/000483d2-8b2d-4403-9652-d4d86d55f7b5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-bcsmj\" (UID: \"000483d2-8b2d-4403-9652-d4d86d55f7b5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.242716 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsw92\" (UniqueName: \"kubernetes.io/projected/fb7c802e-5175-4b0b-8b70-91efb1c83fa1-kube-api-access-nsw92\") pod \"package-server-manager-789f6589d5-46sg8\" (UID: \"fb7c802e-5175-4b0b-8b70-91efb1c83fa1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.247229 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.247389 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f77016a6-cd35-49df-b6b5-65b4858b41c9-secret-volume\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.269206 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.273936 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.277426 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vg4v\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-kube-api-access-2vg4v\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.289987 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.291140 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng7x4\" (UniqueName: \"kubernetes.io/projected/40656b16-31ff-4065-b87c-5678308e9fb4-kube-api-access-ng7x4\") pod \"service-ca-operator-777779d784-tdhsc\" (UID: \"40656b16-31ff-4065-b87c-5678308e9fb4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.292132 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.293586 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.293773 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:24.793756464 +0000 UTC m=+153.768251284 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.293799 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pd7s\" (UniqueName: \"kubernetes.io/projected/9dbe3199-893b-4d67-a447-fa3223c321e3-kube-api-access-9pd7s\") pod \"ingress-canary-bp2vx\" (UID: \"9dbe3199-893b-4d67-a447-fa3223c321e3\") " pod="openshift-ingress-canary/ingress-canary-bp2vx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.293831 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-node-bootstrap-token\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.293867 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv78n\" (UniqueName: \"kubernetes.io/projected/06e36dcf-9973-4369-8dc0-735cc42a68aa-kube-api-access-qv78n\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.293886 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-csi-data-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294643 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06e36dcf-9973-4369-8dc0-735cc42a68aa-config-volume\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294678 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/06e36dcf-9973-4369-8dc0-735cc42a68aa-metrics-tls\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294715 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-certs\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294739 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294767 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-socket-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294786 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtrdt\" (UniqueName: \"kubernetes.io/projected/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-kube-api-access-qtrdt\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294819 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-plugins-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294845 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-mountpoint-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294878 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9dbe3199-893b-4d67-a447-fa3223c321e3-cert\") pod \"ingress-canary-bp2vx\" (UID: \"9dbe3199-893b-4d67-a447-fa3223c321e3\") " pod="openshift-ingress-canary/ingress-canary-bp2vx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294900 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-registration-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294920 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5jbw\" (UniqueName: \"kubernetes.io/projected/13b4eba0-42e7-448a-998c-d7aaeffc9218-kube-api-access-v5jbw\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.294679 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qtlw7"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.296849 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-plugins-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.297236 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:24.797217199 +0000 UTC m=+153.771712019 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.297312 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-mountpoint-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.297448 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-socket-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.297455 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-csi-data-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.297496 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/13b4eba0-42e7-448a-998c-d7aaeffc9218-registration-dir\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.298052 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06e36dcf-9973-4369-8dc0-735cc42a68aa-config-volume\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.298228 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-certs\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.299653 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qgskm\" (UID: \"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.299788 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/06e36dcf-9973-4369-8dc0-735cc42a68aa-metrics-tls\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.300718 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-node-bootstrap-token\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.311519 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9dbe3199-893b-4d67-a447-fa3223c321e3-cert\") pod \"ingress-canary-bp2vx\" (UID: \"9dbe3199-893b-4d67-a447-fa3223c321e3\") " pod="openshift-ingress-canary/ingress-canary-bp2vx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.315307 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj9kq\" (UniqueName: \"kubernetes.io/projected/4b3348f9-d9b9-4efb-9eb8-877f58b0d59c-kube-api-access-fj9kq\") pod \"router-default-5444994796-sc5pm\" (UID: \"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c\") " pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.352318 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svs7x\" (UniqueName: \"kubernetes.io/projected/a102d1fa-69a5-4826-ac40-e07ab4087558-kube-api-access-svs7x\") pod \"packageserver-d55dfcdfc-dkdgm\" (UID: \"a102d1fa-69a5-4826-ac40-e07ab4087558\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.371095 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-bound-sa-token\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.390495 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99fafd36-91d7-45b8-a321-59307d79e5ce-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5kfk4\" (UID: \"99fafd36-91d7-45b8-a321-59307d79e5ce\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.396066 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.396446 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:24.896430981 +0000 UTC m=+153.870925801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.401702 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjwcj\" (UniqueName: \"kubernetes.io/projected/308389f1-6962-45be-87b0-1b061caa2c50-kube-api-access-vjwcj\") pod \"service-ca-9c57cc56f-rtnmx\" (UID: \"308389f1-6962-45be-87b0-1b061caa2c50\") " pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.432044 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ql4r\" (UniqueName: \"kubernetes.io/projected/f77016a6-cd35-49df-b6b5-65b4858b41c9-kube-api-access-2ql4r\") pod \"collect-profiles-29526240-lwsqk\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.444571 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55q6w\" (UniqueName: \"kubernetes.io/projected/95a270e5-7a08-4238-a65d-96986f2d2c39-kube-api-access-55q6w\") pod \"marketplace-operator-79b997595-sjv8t\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.445382 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.458893 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.472241 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npgsj\" (UniqueName: \"kubernetes.io/projected/000483d2-8b2d-4403-9652-d4d86d55f7b5-kube-api-access-npgsj\") pod \"multus-admission-controller-857f4d67dd-bcsmj\" (UID: \"000483d2-8b2d-4403-9652-d4d86d55f7b5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.495882 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmt4n\" (UniqueName: \"kubernetes.io/projected/912a3267-581f-47b1-873c-b0e4d8dd3768-kube-api-access-zmt4n\") pod \"catalog-operator-68c6474976-dgw2h\" (UID: \"912a3267-581f-47b1-873c-b0e4d8dd3768\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.497248 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.497720 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:24.997704094 +0000 UTC m=+153.972198914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.507842 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.515086 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.515392 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.526237 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.540188 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv78n\" (UniqueName: \"kubernetes.io/projected/06e36dcf-9973-4369-8dc0-735cc42a68aa-kube-api-access-qv78n\") pod \"dns-default-9ls2l\" (UID: \"06e36dcf-9973-4369-8dc0-735cc42a68aa\") " pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.541159 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.543935 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.549060 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.560076 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pd7s\" (UniqueName: \"kubernetes.io/projected/9dbe3199-893b-4d67-a447-fa3223c321e3-kube-api-access-9pd7s\") pod \"ingress-canary-bp2vx\" (UID: \"9dbe3199-893b-4d67-a447-fa3223c321e3\") " pod="openshift-ingress-canary/ingress-canary-bp2vx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.582939 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5jbw\" (UniqueName: \"kubernetes.io/projected/13b4eba0-42e7-448a-998c-d7aaeffc9218-kube-api-access-v5jbw\") pod \"csi-hostpathplugin-trh7g\" (UID: \"13b4eba0-42e7-448a-998c-d7aaeffc9218\") " pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.599578 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.599908 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.09989068 +0000 UTC m=+154.074385500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.601898 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.612653 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtrdt\" (UniqueName: \"kubernetes.io/projected/bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156-kube-api-access-qtrdt\") pod \"machine-config-server-gd69n\" (UID: \"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156\") " pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.613979 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.626348 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.638314 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.652409 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.652667 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-bp2vx" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.653642 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-k8zcr"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.655327 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.657163 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-44l7j"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.698169 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-trh7g" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.699836 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-gd69n" Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.703790 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.704428 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.204408683 +0000 UTC m=+154.178903503 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.745847 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-xjbrz"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.777944 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg"] Feb 20 08:08:24 crc kubenswrapper[4948]: W0220 08:08:24.785486 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c070b67_c2ca_459b_a1b7_813a8833e27e.slice/crio-f24acc5a18bb988f4e5c8c67e817d3ea2636025712e6b3477a09950d4595e36b WatchSource:0}: Error finding container f24acc5a18bb988f4e5c8c67e817d3ea2636025712e6b3477a09950d4595e36b: Status 404 returned error can't find the container with id f24acc5a18bb988f4e5c8c67e817d3ea2636025712e6b3477a09950d4595e36b Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.788895 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" event={"ID":"de9b95d8-fa65-438c-bef9-2aba46044ec1","Type":"ContainerStarted","Data":"80bb30586955a2bb2425be91e482784aaca7deedc011442d5ba01e04abd2835b"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.795943 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fhfhr"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.804647 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.805897 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.305865902 +0000 UTC m=+154.280360722 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.811312 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" event={"ID":"2c59c4e2-97ae-494a-b33a-a542266cf233","Type":"ContainerStarted","Data":"46dc1f39b430dec05ad46ae2a5d1777f0c144a8a7b1211db5a8b9a66412709c3"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.827236 4948 generic.go:334] "Generic (PLEG): container finished" podID="c7c6bf9e-0846-4391-9126-a00d68b2e627" containerID="2b2020a1e307e87d92e1c75ed8ea2540d79082ae7f10ec4114f5d6a47e553948" exitCode=0 Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.827425 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" event={"ID":"c7c6bf9e-0846-4391-9126-a00d68b2e627","Type":"ContainerDied","Data":"2b2020a1e307e87d92e1c75ed8ea2540d79082ae7f10ec4114f5d6a47e553948"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.833169 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" event={"ID":"5bc3d0cd-eacd-4b25-9acd-853e49db7b47","Type":"ContainerStarted","Data":"c3ef45d9487fa7c7fcc3b038b4c27102a6491a3966150aa1fc76eec4719774ea"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.833213 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" event={"ID":"5bc3d0cd-eacd-4b25-9acd-853e49db7b47","Type":"ContainerStarted","Data":"5a21d2c4a35b9613a1f1d569beddc744e500d407a9692ba5a38e15f50751b685"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.843475 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" event={"ID":"fabbac48-e5db-4057-b9a5-69118c38c667","Type":"ContainerStarted","Data":"3cc6318d369798be51a8e3af59f16c5cf66e1273c8642f9c640bdcaf2fbdb905"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.843519 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" event={"ID":"fabbac48-e5db-4057-b9a5-69118c38c667","Type":"ContainerStarted","Data":"0078d6f35d4b09153ac5f4519a90739d3b7c1a3332b872561f2b13e7ca2d48f9"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.854233 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" event={"ID":"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb","Type":"ContainerStarted","Data":"e6bd43e85d4f40780c71d206d3cda54da562edaaf59358c74e9a9f95f771792a"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.873145 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" event={"ID":"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72","Type":"ContainerStarted","Data":"45cb715b9778f0421e1f7a41623f8cc620132945850c903c2ad1e537a17ac52f"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.873193 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" event={"ID":"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72","Type":"ContainerStarted","Data":"e8e9ca0e8794f3f1b33e90fba54ddb0b40a7135497207e0f2a0473343f74e227"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.876123 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-grnfc"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.891999 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" event={"ID":"3f4bc9ec-3eda-4d56-9696-df15c14d2e4a","Type":"ContainerStarted","Data":"6a636a454976750e16ccc1469c63f538a09356f52d37fa7566be8b50b72c9709"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.907251 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" event={"ID":"01d95eeb-e421-4fe2-a24a-44c3014f1f6e","Type":"ContainerStarted","Data":"7635ec0ec7c24c8203327d8beda70ffb895aa8078a28ac99baf419a8966811ca"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.907299 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" event={"ID":"01d95eeb-e421-4fe2-a24a-44c3014f1f6e","Type":"ContainerStarted","Data":"fd9df30c5d63b531a8098b8313e0aead45e05e65f20b2d3a9242e2f927d8db1f"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.908349 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:24 crc kubenswrapper[4948]: E0220 08:08:24.908620 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.408608771 +0000 UTC m=+154.383103591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.909125 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sjnxm"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.917275 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-x8fgj"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.942252 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" event={"ID":"e029f4ab-6a61-4fcf-9997-b51c1cbe5675","Type":"ContainerStarted","Data":"989ca99ae7d68ec8b3478c17b0e5e440952effaa6b8e84cc390e63652c262e43"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.947177 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" event={"ID":"b1099221-b48f-4756-b133-0fed16b1e225","Type":"ContainerStarted","Data":"73250930b110d9ec15accb42a0a2a49b5ad06046d887c0bd11c8d587d39d5ab1"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.968475 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" event={"ID":"a149fe37-c748-4120-9116-1da4b680d880","Type":"ContainerStarted","Data":"c6a58e0e094dca5eb92e842406fb7bb59ef92c395eadd8ac868c827c07eb9392"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.968521 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" event={"ID":"a149fe37-c748-4120-9116-1da4b680d880","Type":"ContainerStarted","Data":"64d960029fc420361a87e4a17c366841af210e237e40e75ca94a40969f226b89"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.982515 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" event={"ID":"cf612779-d96d-48f1-9958-5f92ac6e3d40","Type":"ContainerStarted","Data":"d785a33073edb1bc2cec1d3eb6855e83ab9c2dd8f8168b35d03bdcffc0a1f5da"} Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.986965 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.993655 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc"] Feb 20 08:08:24 crc kubenswrapper[4948]: I0220 08:08:24.996541 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" event={"ID":"02036666-815d-4282-86f0-1b4163ea7cf9","Type":"ContainerStarted","Data":"b835c835e64d80d465d81fe9f329df642a1182ff71ea166fabeb1ad480fae83b"} Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:24.996582 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" event={"ID":"02036666-815d-4282-86f0-1b4163ea7cf9","Type":"ContainerStarted","Data":"42bd08c72bddb366038d9580429774e95c2763c7272273dc2cac5528d750116b"} Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.001811 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm"] Feb 20 08:08:25 crc kubenswrapper[4948]: W0220 08:08:25.005119 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc13d9fe_fd58_4d3b_9278_4f3e59da5976.slice/crio-1e15ef1a019fef26569179f0c621d2a2878c8686434cf1cb9b657cedc5b91011 WatchSource:0}: Error finding container 1e15ef1a019fef26569179f0c621d2a2878c8686434cf1cb9b657cedc5b91011: Status 404 returned error can't find the container with id 1e15ef1a019fef26569179f0c621d2a2878c8686434cf1cb9b657cedc5b91011 Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.009830 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.010141 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.51011534 +0000 UTC m=+154.484610160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.010313 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.011072 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.511061514 +0000 UTC m=+154.485556334 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.018535 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.111220 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.112354 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.612336137 +0000 UTC m=+154.586830957 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.167563 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.201220 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.219180 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.219615 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.719599018 +0000 UTC m=+154.694093838 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.258943 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-trh7g"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.320679 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.321215 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.821195599 +0000 UTC m=+154.795690419 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.430589 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.432355 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:25.932335285 +0000 UTC m=+154.906830095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.552092 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.552585 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.052567503 +0000 UTC m=+155.027062323 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.648005 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-thcgl" podStartSLOduration=131.647952043 podStartE2EDuration="2m11.647952043s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:25.636908412 +0000 UTC m=+154.611403232" watchObservedRunningTime="2026-02-20 08:08:25.647952043 +0000 UTC m=+154.622446863" Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.660369 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.660836 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.160822548 +0000 UTC m=+155.135317368 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.763051 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.763150 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.263119787 +0000 UTC m=+155.237614597 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.763685 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.764207 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.264195363 +0000 UTC m=+155.238690183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.815539 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-9ls2l"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.817571 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.827498 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.837906 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.871512 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.872086 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.372069489 +0000 UTC m=+155.346564309 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.930728 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-bcsmj"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.931896 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sjv8t"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.959431 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rtnmx"] Feb 20 08:08:25 crc kubenswrapper[4948]: I0220 08:08:25.973265 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:25 crc kubenswrapper[4948]: E0220 08:08:25.973659 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.47364294 +0000 UTC m=+155.448137750 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:25.997509 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" podStartSLOduration=131.997488525 podStartE2EDuration="2m11.997488525s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:25.996112251 +0000 UTC m=+154.970607071" watchObservedRunningTime="2026-02-20 08:08:25.997488525 +0000 UTC m=+154.971983345" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.075661 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.076709 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.576689277 +0000 UTC m=+155.551184097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.087057 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" event={"ID":"c38c7edf-ca04-4ea4-b15f-d83dea3f545e","Type":"ContainerStarted","Data":"c95f2889704f670ef930f1019a8454e6d7e06bf638a37e204f5949de0826f43c"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.087188 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" event={"ID":"c38c7edf-ca04-4ea4-b15f-d83dea3f545e","Type":"ContainerStarted","Data":"ea691caa5c1656ad51f3d8af407e41e98d60cd663139e39b82d64f3ca6e94e94"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.100519 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-bp2vx"] Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.101236 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" event={"ID":"912a3267-581f-47b1-873c-b0e4d8dd3768","Type":"ContainerStarted","Data":"f0dd589c2408c422566835987233c145aeb400052dad0180e547063e2d0d6740"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.104213 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" podStartSLOduration=132.104197121 podStartE2EDuration="2m12.104197121s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.101039094 +0000 UTC m=+155.075533914" watchObservedRunningTime="2026-02-20 08:08:26.104197121 +0000 UTC m=+155.078691961" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.134443 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rrhs" event={"ID":"1eaf5236-6b0b-48fb-91b6-d9c93dfb8a72","Type":"ContainerStarted","Data":"4575bc4762444482fba32a2a68c64a3d1de4d3e3bb3854c702bed016ecb3b763"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.142542 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" event={"ID":"6c070b67-c2ca-459b-a1b7-813a8833e27e","Type":"ContainerStarted","Data":"27c38be8b9e4ccb4d5b9ff6afe53e0b0eb4b257a4accae7cd8daf9b3703d5ce8"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.142597 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" event={"ID":"6c070b67-c2ca-459b-a1b7-813a8833e27e","Type":"ContainerStarted","Data":"f24acc5a18bb988f4e5c8c67e817d3ea2636025712e6b3477a09950d4595e36b"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.143754 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.164721 4948 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-qljmk container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.164787 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" podUID="6c070b67-c2ca-459b-a1b7-813a8833e27e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.177090 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.177425 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.677411927 +0000 UTC m=+155.651906747 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.190253 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-grnfc" event={"ID":"bc13d9fe-fd58-4d3b-9278-4f3e59da5976","Type":"ContainerStarted","Data":"1e15ef1a019fef26569179f0c621d2a2878c8686434cf1cb9b657cedc5b91011"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.206344 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" event={"ID":"05fe2d3b-0a53-4bbe-b0aa-be07acb159bb","Type":"ContainerStarted","Data":"e08a5f2d9b0423afb18692d9ad335a38e239c46dd9ffe8714f61e3f2a149ddf6"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.223822 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" event={"ID":"40656b16-31ff-4065-b87c-5678308e9fb4","Type":"ContainerStarted","Data":"a032e77d70839df518e603667ef479b401e6f48da056e73539ba49efbcdbcc37"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.243429 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" event={"ID":"de9b95d8-fa65-438c-bef9-2aba46044ec1","Type":"ContainerStarted","Data":"e7637483275f788126b9aeb5d1bb4c5cdba980ce676e8999b4f7749cf3b9f849"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.243488 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" event={"ID":"de9b95d8-fa65-438c-bef9-2aba46044ec1","Type":"ContainerStarted","Data":"90e4321361a39876f68035a8a153661f0dc18b7c4540e6d751217c691e0141b7"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.245631 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bgm85" podStartSLOduration=132.245612709 podStartE2EDuration="2m12.245612709s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.243234721 +0000 UTC m=+155.217729541" watchObservedRunningTime="2026-02-20 08:08:26.245612709 +0000 UTC m=+155.220107529" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.273386 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k8zcr" event={"ID":"f81366e5-fbdb-4a0b-a91f-f603d9b8570d","Type":"ContainerStarted","Data":"61712307e24396811f0211771ba686b9bbc2eb0844719329e63b16d44d321de6"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.278576 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.279898 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.77988045 +0000 UTC m=+155.754375270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.281112 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" event={"ID":"f77016a6-cd35-49df-b6b5-65b4858b41c9","Type":"ContainerStarted","Data":"92efa091cdddd802dc78aa2d9bf128d9d0f0ac869c640718f1c1a8d1aead053f"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.345646 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" event={"ID":"b1099221-b48f-4756-b133-0fed16b1e225","Type":"ContainerStarted","Data":"68e683b1d4b19ead7db68fed68ae56db80880a0e972aa74ee98c29b93213ec23"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.364298 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-gd69n" event={"ID":"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156","Type":"ContainerStarted","Data":"7afe19488bbf8afd4db3d1cdf3138ac934b18ee11bcc557b188b23131293e2b0"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.381443 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.383243 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.883225324 +0000 UTC m=+155.857720144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.404462 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" event={"ID":"3f4bc9ec-3eda-4d56-9696-df15c14d2e4a","Type":"ContainerStarted","Data":"27e69ab691f0c40ba85f35afadd1e94a256245e171d5c78e3edbb1d1daa7626d"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.431276 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" event={"ID":"6f1e0caa-03ee-4025-9f2a-6941090c178b","Type":"ContainerStarted","Data":"cabbaec805df0b1917ae1dca77b867f02165355a9ddd40f61184fbc442797a01"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.466845 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" event={"ID":"cf612779-d96d-48f1-9958-5f92ac6e3d40","Type":"ContainerStarted","Data":"36f337a8a94ffe20a346a9ea0395f661eb311a188ba455aa9e3379d139dafd83"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.482765 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.482938 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" event={"ID":"5bc3d0cd-eacd-4b25-9acd-853e49db7b47","Type":"ContainerStarted","Data":"10ec9cf1938d190ec36da83cd5d3e21b002f9026662192a8c0885233df749235"} Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.483122 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:26.983105724 +0000 UTC m=+155.957600544 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.523209 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" event={"ID":"02036666-815d-4282-86f0-1b4163ea7cf9","Type":"ContainerStarted","Data":"173d3953bfe8097f250ffa8ab48fb26722e29618fcdb6e7c03f5c33121453669"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.529858 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vtkbz" podStartSLOduration=132.52983364 podStartE2EDuration="2m12.52983364s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.526440416 +0000 UTC m=+155.500935236" watchObservedRunningTime="2026-02-20 08:08:26.52983364 +0000 UTC m=+155.504328460" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.566242 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" event={"ID":"0b98dbb3-7986-475d-8028-0879d2fed2af","Type":"ContainerStarted","Data":"1f5342e62238c49aaaf9cc95280bea091a38ab7b2dccb6e71f5b3c95eea48e86"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.582781 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-trh7g" event={"ID":"13b4eba0-42e7-448a-998c-d7aaeffc9218","Type":"ContainerStarted","Data":"48fc2a7997571c36fba0d9d88c82b7bc0153d8fb89341ddef42a5a815a193a0a"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.584505 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.584870 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.084855889 +0000 UTC m=+156.059350709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.591061 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" event={"ID":"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf","Type":"ContainerStarted","Data":"74fbc575b905b01e85f8fd524cf1ed8beb2153e8f1ad32af0cdb433f5f045a3f"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.592641 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" event={"ID":"21db7edd-6db3-4afa-b470-0016a9c5afb9","Type":"ContainerStarted","Data":"84341e8559dd79e093cb4de6142b9a22ac6a50040f68f50623fb6ec98aa5e33a"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.596841 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.643155 4948 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-fhfhr container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.643335 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" podUID="21db7edd-6db3-4afa-b470-0016a9c5afb9" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.643683 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nzj8d" event={"ID":"e029f4ab-6a61-4fcf-9997-b51c1cbe5675","Type":"ContainerStarted","Data":"787a481b6915374903830553ffbff88aa0c720e581c4c467bfca9610de775aa3"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.686270 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" event={"ID":"df73228e-df1f-4e9c-aafe-493bb743f98a","Type":"ContainerStarted","Data":"76dafdb5407bf70db2d0cf3c70d0d7427c429ce3ab8c2b2e1a871cb56382ee9a"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.686573 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.686630 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.186614695 +0000 UTC m=+156.161109515 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.687773 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.689479 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.189461884 +0000 UTC m=+156.163956704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.691098 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" event={"ID":"fb7c802e-5175-4b0b-8b70-91efb1c83fa1","Type":"ContainerStarted","Data":"c4e3e88d81b8869b53780024210a6a2d4dc27ab82243c5eadb9dbdba0329ef31"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.701117 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" event={"ID":"0b81d26a-f92a-4602-89ef-5dd9fb24a32f","Type":"ContainerStarted","Data":"47f1a06008e2047099df4b7d2d4f0a3a1cddadec6a353f1e8b9d66189bc5b84f"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.705786 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9ls2l" event={"ID":"06e36dcf-9973-4369-8dc0-735cc42a68aa","Type":"ContainerStarted","Data":"2eea3de6acb8faeaefa1a5be0833ddd962a2c40a7011080ed01a67a8d98e108c"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.712917 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" event={"ID":"e9f46a23-4535-4d64-ae76-065435f8f762","Type":"ContainerStarted","Data":"616e47f1bb6448f32a0fe292b91640bb6d73df61cbdba5efe885de734cecb342"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.726665 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" podStartSLOduration=132.726557214 podStartE2EDuration="2m12.726557214s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.70560197 +0000 UTC m=+155.680096790" watchObservedRunningTime="2026-02-20 08:08:26.726557214 +0000 UTC m=+155.701052034" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.737364 4948 csr.go:261] certificate signing request csr-5cvr7 is approved, waiting to be issued Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.750269 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-xjbrz" event={"ID":"bbc3437e-8729-46f8-aee2-ad3a4679097f","Type":"ContainerStarted","Data":"c478364bdabada6a44416428ef78a5f89ef672caf5d301b485497a22b89e761e"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.751056 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.751327 4948 csr.go:257] certificate signing request csr-5cvr7 is issued Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.761462 4948 patch_prober.go:28] interesting pod/console-operator-58897d9998-xjbrz container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.761532 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-xjbrz" podUID="bbc3437e-8729-46f8-aee2-ad3a4679097f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.763633 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7x85x" podStartSLOduration=132.763611563 podStartE2EDuration="2m12.763611563s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.741321206 +0000 UTC m=+155.715816026" watchObservedRunningTime="2026-02-20 08:08:26.763611563 +0000 UTC m=+155.738106373" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.766364 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-sc5pm" event={"ID":"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c","Type":"ContainerStarted","Data":"ad705231484db9909f86eb5c84a06a81f6e5a0c076e43683547b25a3a95dbe85"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.770690 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-fthdq" podStartSLOduration=132.770674156 podStartE2EDuration="2m12.770674156s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.766704559 +0000 UTC m=+155.741199369" watchObservedRunningTime="2026-02-20 08:08:26.770674156 +0000 UTC m=+155.745168976" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.790702 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.791994 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.291960088 +0000 UTC m=+156.266454898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.792342 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" event={"ID":"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573","Type":"ContainerStarted","Data":"dc3f8115d345f257d8228dd6508393b32ec4c3016cf8750d09de9a72aa5aa83e"} Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.823763 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4kfs" podStartSLOduration=132.823731757 podStartE2EDuration="2m12.823731757s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.80630979 +0000 UTC m=+155.780804610" watchObservedRunningTime="2026-02-20 08:08:26.823731757 +0000 UTC m=+155.798226577" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.836108 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-qtlw7" podStartSLOduration=132.8360877 podStartE2EDuration="2m12.8360877s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.826639519 +0000 UTC m=+155.801134339" watchObservedRunningTime="2026-02-20 08:08:26.8360877 +0000 UTC m=+155.810582520" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.858831 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-fl47j" podStartSLOduration=132.858800157 podStartE2EDuration="2m12.858800157s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.857438444 +0000 UTC m=+155.831933264" watchObservedRunningTime="2026-02-20 08:08:26.858800157 +0000 UTC m=+155.833294977" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.876103 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" podStartSLOduration=132.876083151 podStartE2EDuration="2m12.876083151s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.875495777 +0000 UTC m=+155.849990597" watchObservedRunningTime="2026-02-20 08:08:26.876083151 +0000 UTC m=+155.850577961" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.892761 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.893910 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.393895158 +0000 UTC m=+156.368389978 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.923942 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-qfgtj" podStartSLOduration=132.923916304 podStartE2EDuration="2m12.923916304s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:26.920587582 +0000 UTC m=+155.895082402" watchObservedRunningTime="2026-02-20 08:08:26.923916304 +0000 UTC m=+155.898411124" Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.995450 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.995773 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.495736205 +0000 UTC m=+156.470231025 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:26 crc kubenswrapper[4948]: I0220 08:08:26.996344 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:26 crc kubenswrapper[4948]: E0220 08:08:26.996867 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.496850763 +0000 UTC m=+156.471345583 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.003033 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" podStartSLOduration=133.003009834 podStartE2EDuration="2m13.003009834s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:27.00244306 +0000 UTC m=+155.976937870" watchObservedRunningTime="2026-02-20 08:08:27.003009834 +0000 UTC m=+155.977504654" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.102505 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.102713 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.602697388 +0000 UTC m=+156.577192208 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.103171 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.103530 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.603522989 +0000 UTC m=+156.578017809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.105875 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-sc5pm" podStartSLOduration=133.105851556 podStartE2EDuration="2m13.105851556s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:27.101649983 +0000 UTC m=+156.076144803" watchObservedRunningTime="2026-02-20 08:08:27.105851556 +0000 UTC m=+156.080346376" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.108088 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-xjbrz" podStartSLOduration=133.108079771 podStartE2EDuration="2m13.108079771s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:27.076721631 +0000 UTC m=+156.051216451" watchObservedRunningTime="2026-02-20 08:08:27.108079771 +0000 UTC m=+156.082574591" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.206437 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.207481 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.707449297 +0000 UTC m=+156.681944117 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.308209 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.308655 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.808638299 +0000 UTC m=+156.783133119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.409128 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.909109683 +0000 UTC m=+156.883604503 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.409038 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.409377 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.409724 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:27.909713558 +0000 UTC m=+156.884208378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.463058 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.463454 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.463482 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.512855 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.513866 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.013846831 +0000 UTC m=+156.988341651 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.615102 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.615489 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.115475764 +0000 UTC m=+157.089970584 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.717203 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.717554 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.217523356 +0000 UTC m=+157.192018176 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.719192 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.719687 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.219670649 +0000 UTC m=+157.194165469 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.757062 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-20 08:03:26 +0000 UTC, rotation deadline is 2026-12-18 03:05:35.800339458 +0000 UTC Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.757412 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7218h57m8.042930834s for next certificate rotation Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.801827 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" event={"ID":"40656b16-31ff-4065-b87c-5678308e9fb4","Type":"ContainerStarted","Data":"8010bfb6dfead3b5a6d77b49e41e950061de92d8c07b3666d38680482a3844af"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.808908 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" event={"ID":"000483d2-8b2d-4403-9652-d4d86d55f7b5","Type":"ContainerStarted","Data":"8142fe9f926bd6c4b6cdd3b8e3dc4f48366e5d712648e9b3e5e5813d7095a93c"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.808956 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" event={"ID":"000483d2-8b2d-4403-9652-d4d86d55f7b5","Type":"ContainerStarted","Data":"e610ee04d021213df53cb4b947552b128f5bd097bb682e7b6347c12abee71f0f"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.820784 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.822228 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.322203584 +0000 UTC m=+157.296698394 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.825024 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tdhsc" podStartSLOduration=133.824998792 podStartE2EDuration="2m13.824998792s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:27.824417948 +0000 UTC m=+156.798912768" watchObservedRunningTime="2026-02-20 08:08:27.824998792 +0000 UTC m=+156.799493612" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.830052 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" event={"ID":"df73228e-df1f-4e9c-aafe-493bb743f98a","Type":"ContainerStarted","Data":"f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.831153 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.843718 4948 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-sjnxm container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.33:6443/healthz\": dial tcp 10.217.0.33:6443: connect: connection refused" start-of-body= Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.844060 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" podUID="df73228e-df1f-4e9c-aafe-493bb743f98a" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.33:6443/healthz\": dial tcp 10.217.0.33:6443: connect: connection refused" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.853427 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" event={"ID":"e5e9ef3a-e3aa-43ed-93d6-2b854f716ddf","Type":"ContainerStarted","Data":"158588997f0116bf9c80751e42eec3bc450832a51470d5e36f65ed2c94dbeb58"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.866051 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" podStartSLOduration=133.866019898 podStartE2EDuration="2m13.866019898s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:27.858192406 +0000 UTC m=+156.832687226" watchObservedRunningTime="2026-02-20 08:08:27.866019898 +0000 UTC m=+156.840514718" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.874206 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" event={"ID":"fb7c802e-5175-4b0b-8b70-91efb1c83fa1","Type":"ContainerStarted","Data":"c3dce3a4ba3c8fc0032fff443af8742149cd5cdf42877e90386f11470aa73932"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.874258 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" event={"ID":"fb7c802e-5175-4b0b-8b70-91efb1c83fa1","Type":"ContainerStarted","Data":"25c94662e3decbb3b4e2aaba9a0cb2773839779bffa1f9f9a23009b452d4dfcd"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.874941 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.910481 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-bp2vx" event={"ID":"9dbe3199-893b-4d67-a447-fa3223c321e3","Type":"ContainerStarted","Data":"92a8b354f092998d33eb260a3546688c2937af73a1b2ae389b0bea11d8f50794"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.914610 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-bp2vx" event={"ID":"9dbe3199-893b-4d67-a447-fa3223c321e3","Type":"ContainerStarted","Data":"63e9556a2a22b7fd0f1d180db3021b4e14c73e783441d8d43c988d3dd1c01136"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.922050 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:27 crc kubenswrapper[4948]: E0220 08:08:27.926443 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.426415838 +0000 UTC m=+157.400910648 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.944991 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qgskm" podStartSLOduration=133.944947363 podStartE2EDuration="2m13.944947363s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:27.88813299 +0000 UTC m=+156.862627810" watchObservedRunningTime="2026-02-20 08:08:27.944947363 +0000 UTC m=+156.919442183" Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.968377 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" event={"ID":"b1099221-b48f-4756-b133-0fed16b1e225","Type":"ContainerStarted","Data":"f602ad97a5fc08e4d185a5a9671744a590f1170b60b5ca4802d6f917630253cf"} Feb 20 08:08:27 crc kubenswrapper[4948]: I0220 08:08:27.978230 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" podStartSLOduration=133.978210818 podStartE2EDuration="2m13.978210818s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:27.947505005 +0000 UTC m=+156.921999825" watchObservedRunningTime="2026-02-20 08:08:27.978210818 +0000 UTC m=+156.952705638" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.012807 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-bp2vx" podStartSLOduration=7.012784476 podStartE2EDuration="7.012784476s" podCreationTimestamp="2026-02-20 08:08:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:27.978958287 +0000 UTC m=+156.953453117" watchObservedRunningTime="2026-02-20 08:08:28.012784476 +0000 UTC m=+156.987279296" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.013912 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lrdf2" podStartSLOduration=134.013907224 podStartE2EDuration="2m14.013907224s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.01088422 +0000 UTC m=+156.985379040" watchObservedRunningTime="2026-02-20 08:08:28.013907224 +0000 UTC m=+156.988402034" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.014572 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-sc5pm" event={"ID":"4b3348f9-d9b9-4efb-9eb8-877f58b0d59c","Type":"ContainerStarted","Data":"c749eafa49326acbaefed5b7bdb334d8f07a133966a45121fbd7300edd8f537b"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.031473 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.037302 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" event={"ID":"f77016a6-cd35-49df-b6b5-65b4858b41c9","Type":"ContainerStarted","Data":"0520ff9e3a8f12a5228a352b59373978c12d5f10bb51923b12eab15acf564086"} Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.038399 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.538379204 +0000 UTC m=+157.512874024 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.059290 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-xjbrz" event={"ID":"bbc3437e-8729-46f8-aee2-ad3a4679097f","Type":"ContainerStarted","Data":"e0124670c1008f3fd7929e53fe9680fae097c86e551ab7d64a8f6733d6449e84"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.070025 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" podStartSLOduration=134.07000051 podStartE2EDuration="2m14.07000051s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.068913163 +0000 UTC m=+157.043407983" watchObservedRunningTime="2026-02-20 08:08:28.07000051 +0000 UTC m=+157.044495330" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.081345 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-klsxg" event={"ID":"0b81d26a-f92a-4602-89ef-5dd9fb24a32f","Type":"ContainerStarted","Data":"b029d80c1e577342eaa2888ffe51427b6da3e30eb7f1e2bafc7ab86e988b66c1"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.092683 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" event={"ID":"c7b67dd6-4025-45ad-98f8-2e7f8ab4b573","Type":"ContainerStarted","Data":"f08b11d190142a1c0ac494119229274702a7dd496184ab826a44da91c5aeb639"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.093350 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.103376 4948 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6zdkm container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.103438 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" podUID="c7b67dd6-4025-45ad-98f8-2e7f8ab4b573" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.115271 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" event={"ID":"308389f1-6962-45be-87b0-1b061caa2c50","Type":"ContainerStarted","Data":"0485996e9a881ae3c35cf440e5293526ff386a26f8480ffec43777bc9f6cd2cc"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.115475 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" event={"ID":"308389f1-6962-45be-87b0-1b061caa2c50","Type":"ContainerStarted","Data":"406d80af236a11af7cfd4a41db7671d2067912bafd6e7cebcc426a1590e948e3"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.135448 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.136406 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.636384217 +0000 UTC m=+157.610879107 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.149624 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" event={"ID":"95a270e5-7a08-4238-a65d-96986f2d2c39","Type":"ContainerStarted","Data":"15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.149946 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" event={"ID":"95a270e5-7a08-4238-a65d-96986f2d2c39","Type":"ContainerStarted","Data":"936301da215b82b117fb711a926cf169f025d16470e0871be5907e8efa5aec8e"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.151736 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.158487 4948 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-sjv8t container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.159013 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" podUID="95a270e5-7a08-4238-a65d-96986f2d2c39" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.162510 4948 generic.go:334] "Generic (PLEG): container finished" podID="2c59c4e2-97ae-494a-b33a-a542266cf233" containerID="b54927298d346e7bc2198cc485b8d321498aea49b86a1ada6f9600eb7b7e5912" exitCode=0 Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.162708 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" event={"ID":"2c59c4e2-97ae-494a-b33a-a542266cf233","Type":"ContainerDied","Data":"b54927298d346e7bc2198cc485b8d321498aea49b86a1ada6f9600eb7b7e5912"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.177121 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-rtnmx" podStartSLOduration=133.177093826 podStartE2EDuration="2m13.177093826s" podCreationTimestamp="2026-02-20 08:06:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.176518342 +0000 UTC m=+157.151013162" watchObservedRunningTime="2026-02-20 08:08:28.177093826 +0000 UTC m=+157.151588646" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.178552 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" podStartSLOduration=134.178542891 podStartE2EDuration="2m14.178542891s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.128936175 +0000 UTC m=+157.103430995" watchObservedRunningTime="2026-02-20 08:08:28.178542891 +0000 UTC m=+157.153037711" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.189858 4948 generic.go:334] "Generic (PLEG): container finished" podID="0b98dbb3-7986-475d-8028-0879d2fed2af" containerID="ddd4a11d3470308b5ca2eb6c345ef2617f7ff6bae43e4451822a612395bb815e" exitCode=0 Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.189930 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" event={"ID":"0b98dbb3-7986-475d-8028-0879d2fed2af","Type":"ContainerDied","Data":"ddd4a11d3470308b5ca2eb6c345ef2617f7ff6bae43e4451822a612395bb815e"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.219230 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" event={"ID":"21db7edd-6db3-4afa-b470-0016a9c5afb9","Type":"ContainerStarted","Data":"aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.239319 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.239547 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.739513937 +0000 UTC m=+157.714008757 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.241243 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.257582 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.757566079 +0000 UTC m=+157.732060899 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.280789 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" podStartSLOduration=134.280771718 podStartE2EDuration="2m14.280771718s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.206188689 +0000 UTC m=+157.180683509" watchObservedRunningTime="2026-02-20 08:08:28.280771718 +0000 UTC m=+157.255266538" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.292640 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.292766 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" event={"ID":"e9f46a23-4535-4d64-ae76-065435f8f762","Type":"ContainerStarted","Data":"437f16fdb8670c6f1795395ab53f0a3d4b705057dc4b0f5bd948256585c1053e"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.292841 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" event={"ID":"e9f46a23-4535-4d64-ae76-065435f8f762","Type":"ContainerStarted","Data":"2afb989988045404096c867b49d32b1ae96c6879a5066aa9d967cf42bdd1216e"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.331309 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" event={"ID":"6f1e0caa-03ee-4025-9f2a-6941090c178b","Type":"ContainerStarted","Data":"62d64849daf34c87ca026bc1f7a1b40fff2ddf3be9856fd2f487a5a16d63e71b"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.342452 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.342778 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.842732628 +0000 UTC m=+157.817227448 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.343251 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.343804 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.843796044 +0000 UTC m=+157.818290864 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.392499 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" event={"ID":"c7c6bf9e-0846-4391-9126-a00d68b2e627","Type":"ContainerStarted","Data":"a33cdea76478f468d14acad0119d166426114e9269318de6b696bc966c40872c"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.392552 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" event={"ID":"c7c6bf9e-0846-4391-9126-a00d68b2e627","Type":"ContainerStarted","Data":"587dd77e32582f1cad315db93f2588f0ff6e4dc9474fde9a88701bb7760bd9cd"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.409735 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6jlbw" podStartSLOduration=134.409718521 podStartE2EDuration="2m14.409718521s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.371534754 +0000 UTC m=+157.346029564" watchObservedRunningTime="2026-02-20 08:08:28.409718521 +0000 UTC m=+157.384213341" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.413758 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-grnfc" event={"ID":"bc13d9fe-fd58-4d3b-9278-4f3e59da5976","Type":"ContainerStarted","Data":"3d4a3a3e86ff6555077f70891860bf39b098a33875c10bf8c2fe774c9bc00ee1"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.414575 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-grnfc" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.422253 4948 patch_prober.go:28] interesting pod/downloads-7954f5f757-grnfc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.422334 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-grnfc" podUID="bc13d9fe-fd58-4d3b-9278-4f3e59da5976" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.440192 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" event={"ID":"a102d1fa-69a5-4826-ac40-e07ab4087558","Type":"ContainerStarted","Data":"04b654d2c614fb8a9d9b11bb1558405cff6b9ad11400bafa7c2f65d293f00646"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.440241 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" event={"ID":"a102d1fa-69a5-4826-ac40-e07ab4087558","Type":"ContainerStarted","Data":"15a04fb4755610f48f1996977d16d9994621f61b4a6425ddf01ea0f7379ae164"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.441013 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.441898 4948 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-dkdgm container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" start-of-body= Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.441936 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" podUID="a102d1fa-69a5-4826-ac40-e07ab4087558" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.444185 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.445401 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:28.945385405 +0000 UTC m=+157.919880235 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.460155 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" event={"ID":"99fafd36-91d7-45b8-a321-59307d79e5ce","Type":"ContainerStarted","Data":"d2538b6448874d66303a896eee725d42cff670ea7e0b1b3eaf8c102491876a77"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.460640 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" event={"ID":"99fafd36-91d7-45b8-a321-59307d79e5ce","Type":"ContainerStarted","Data":"daae403b300c527e65888b7699ef958c2cc74f2d75fdb04a1659c72e93a624d5"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.472202 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:28 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:28 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:28 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.472261 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.511308 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k8zcr" event={"ID":"f81366e5-fbdb-4a0b-a91f-f603d9b8570d","Type":"ContainerStarted","Data":"3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.546787 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.554173 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" event={"ID":"912a3267-581f-47b1-873c-b0e4d8dd3768","Type":"ContainerStarted","Data":"7bcd605d1ba730936d407eda494a4e9c0c2e556a56b4acbadefc6d4d5fbdf721"} Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.555158 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.055142087 +0000 UTC m=+158.029636907 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.555150 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.563160 4948 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-dgw2h container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.563231 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" podUID="912a3267-581f-47b1-873c-b0e4d8dd3768" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.587345 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" podStartSLOduration=134.587329316 podStartE2EDuration="2m14.587329316s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.584759253 +0000 UTC m=+157.559254063" watchObservedRunningTime="2026-02-20 08:08:28.587329316 +0000 UTC m=+157.561824136" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.602292 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-gd69n" event={"ID":"bd2ba1ed-a8d2-4717-9eeb-4ff450dc3156","Type":"ContainerStarted","Data":"5ccd94cad6ceb1d22526a784585976924a9fad67a354bd99cc01a2852b32de5c"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.639797 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" event={"ID":"3f4bc9ec-3eda-4d56-9696-df15c14d2e4a","Type":"ContainerStarted","Data":"34cf850eb5acf1ade6721905d738cb714924f6cf6b9e924fef7ceb56f03460e0"} Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.651345 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.651897 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.653589 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.153564631 +0000 UTC m=+158.128059451 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.764771 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.766943 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.266928181 +0000 UTC m=+158.241423001 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.852321 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" podStartSLOduration=134.852295714 podStartE2EDuration="2m14.852295714s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.762956723 +0000 UTC m=+157.737451543" watchObservedRunningTime="2026-02-20 08:08:28.852295714 +0000 UTC m=+157.826790534" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.853210 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5kfk4" podStartSLOduration=134.853203137 podStartE2EDuration="2m14.853203137s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.837616914 +0000 UTC m=+157.812111734" watchObservedRunningTime="2026-02-20 08:08:28.853203137 +0000 UTC m=+157.827697957" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.885012 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.885326 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.385309564 +0000 UTC m=+158.359804384 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.979403 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-xjbrz" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.988771 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-grnfc" podStartSLOduration=134.98873926 podStartE2EDuration="2m14.98873926s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.898047376 +0000 UTC m=+157.872542196" watchObservedRunningTime="2026-02-20 08:08:28.98873926 +0000 UTC m=+157.963234080" Feb 20 08:08:28 crc kubenswrapper[4948]: E0220 08:08:28.989144 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.4891312 +0000 UTC m=+158.463626020 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.990313 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" podStartSLOduration=134.990307199 podStartE2EDuration="2m14.990307199s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:28.988511215 +0000 UTC m=+157.963006035" watchObservedRunningTime="2026-02-20 08:08:28.990307199 +0000 UTC m=+157.964802009" Feb 20 08:08:28 crc kubenswrapper[4948]: I0220 08:08:28.988828 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.092717 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.093111 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.59309418 +0000 UTC m=+158.567589000 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.140634 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-gd69n" podStartSLOduration=8.140612855 podStartE2EDuration="8.140612855s" podCreationTimestamp="2026-02-20 08:08:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:29.074254308 +0000 UTC m=+158.048749128" watchObservedRunningTime="2026-02-20 08:08:29.140612855 +0000 UTC m=+158.115107675" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.141510 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-k8zcr" podStartSLOduration=135.141505527 podStartE2EDuration="2m15.141505527s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:29.139209851 +0000 UTC m=+158.113704671" watchObservedRunningTime="2026-02-20 08:08:29.141505527 +0000 UTC m=+158.116000347" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.194636 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.195202 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.695176753 +0000 UTC m=+158.669671763 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.226048 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-cc77j" podStartSLOduration=135.226020559 podStartE2EDuration="2m15.226020559s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:29.225739263 +0000 UTC m=+158.200234083" watchObservedRunningTime="2026-02-20 08:08:29.226020559 +0000 UTC m=+158.200515369" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.295141 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.295448 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.795408601 +0000 UTC m=+158.769903421 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.295523 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.296175 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.796155189 +0000 UTC m=+158.770650009 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.327086 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" podStartSLOduration=135.327064847 podStartE2EDuration="2m15.327064847s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:29.322742411 +0000 UTC m=+158.297237231" watchObservedRunningTime="2026-02-20 08:08:29.327064847 +0000 UTC m=+158.301559667" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.396819 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.397063 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.897032553 +0000 UTC m=+158.871527373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.397121 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.397463 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.897450643 +0000 UTC m=+158.871945463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.466736 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:29 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:29 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:29 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.467349 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.498803 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.499048 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.999010174 +0000 UTC m=+158.973504994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.499106 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.499596 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:29.999587168 +0000 UTC m=+158.974081988 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.599771 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.599911 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.099890318 +0000 UTC m=+159.074385138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.600145 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.600446 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.100437971 +0000 UTC m=+159.074932791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.665075 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9ls2l" event={"ID":"06e36dcf-9973-4369-8dc0-735cc42a68aa","Type":"ContainerStarted","Data":"a3bbdaa9b0b3861669b1c132abe048f8076108b0936f02380e364cfdff3bc95d"} Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.665127 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9ls2l" event={"ID":"06e36dcf-9973-4369-8dc0-735cc42a68aa","Type":"ContainerStarted","Data":"2ed05c7268732ababe3198649757956375bf224a81ac92918b4581897f5a57c1"} Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.665938 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.671353 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" event={"ID":"2c59c4e2-97ae-494a-b33a-a542266cf233","Type":"ContainerStarted","Data":"4e297c40eb98b8058e01dae8aa3d6706d4903ebaed0c20f50c84bc44a9a72a3e"} Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.689748 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" event={"ID":"0b98dbb3-7986-475d-8028-0879d2fed2af","Type":"ContainerStarted","Data":"7ef4f62f031aa5805a790ed6e9c2a2f08a724dd6e1a4e587caa0b73ae425de22"} Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.689794 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.695314 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-9ls2l" podStartSLOduration=8.695297658 podStartE2EDuration="8.695297658s" podCreationTimestamp="2026-02-20 08:08:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:29.69253064 +0000 UTC m=+158.667025450" watchObservedRunningTime="2026-02-20 08:08:29.695297658 +0000 UTC m=+158.669792478" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.701047 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.701496 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.201479679 +0000 UTC m=+159.175974499 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.705190 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-trh7g" event={"ID":"13b4eba0-42e7-448a-998c-d7aaeffc9218","Type":"ContainerStarted","Data":"e6c7793a919930d4f4d795af5b8190d1891a3304b2e5950965a99e5b1db6a306"} Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.705237 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-trh7g" event={"ID":"13b4eba0-42e7-448a-998c-d7aaeffc9218","Type":"ContainerStarted","Data":"9a6fa9fe46f8ef81c7a6d3bc9620483e487ee8d5fa786777059dcee625c46d19"} Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.765138 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" podStartSLOduration=135.76511369 podStartE2EDuration="2m15.76511369s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:29.763552172 +0000 UTC m=+158.738046992" watchObservedRunningTime="2026-02-20 08:08:29.76511369 +0000 UTC m=+158.739608510" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.766113 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" event={"ID":"000483d2-8b2d-4403-9652-d4d86d55f7b5","Type":"ContainerStarted","Data":"5cc9da3f8414ed4e7f3bdc4ca498f288ed0b5722714cb01a26502c4c4a6e3f99"} Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.783916 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-x8fgj" event={"ID":"6f1e0caa-03ee-4025-9f2a-6941090c178b","Type":"ContainerStarted","Data":"47920ed6007ba3c92b551410a6ae9dc8f5f9996533e762ce51d250d6e5f0b84f"} Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.789456 4948 patch_prober.go:28] interesting pod/downloads-7954f5f757-grnfc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.789564 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-grnfc" podUID="bc13d9fe-fd58-4d3b-9278-4f3e59da5976" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.793239 4948 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-sjv8t container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.793294 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" podUID="95a270e5-7a08-4238-a65d-96986f2d2c39" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.800747 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dgw2h" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.801122 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.801987 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.803256 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.303231705 +0000 UTC m=+159.277726695 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.844371 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6zdkm" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.895260 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" podStartSLOduration=135.895244301 podStartE2EDuration="2m15.895244301s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:29.894681707 +0000 UTC m=+158.869176527" watchObservedRunningTime="2026-02-20 08:08:29.895244301 +0000 UTC m=+158.869739121" Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.905495 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.905782 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.405761539 +0000 UTC m=+159.380256359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:29 crc kubenswrapper[4948]: I0220 08:08:29.906397 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:29 crc kubenswrapper[4948]: E0220 08:08:29.934830 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.434802741 +0000 UTC m=+159.409297561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.011800 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.012383 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.512344643 +0000 UTC m=+159.486839463 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.113413 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.113802 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.613787631 +0000 UTC m=+159.588282441 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.186524 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-bcsmj" podStartSLOduration=136.186502674 podStartE2EDuration="2m16.186502674s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:30.182914736 +0000 UTC m=+159.157409556" watchObservedRunningTime="2026-02-20 08:08:30.186502674 +0000 UTC m=+159.160997494" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.214584 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.214760 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.714733776 +0000 UTC m=+159.689228596 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.214865 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.215228 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.715213078 +0000 UTC m=+159.689707898 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.315946 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.316208 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.816168164 +0000 UTC m=+159.790662984 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.316892 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.317404 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.817387474 +0000 UTC m=+159.791882294 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.419415 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.420426 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.92038598 +0000 UTC m=+159.894880800 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.420745 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.421552 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:30.921514937 +0000 UTC m=+159.896009757 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.442715 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nrhl6"] Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.444290 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.446728 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.460657 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nrhl6"] Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.468227 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:30 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:30 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:30 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.468300 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.524682 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.525048 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-catalog-content\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.525129 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc669\" (UniqueName: \"kubernetes.io/projected/f75a50cf-7dae-420b-a00d-671ede3cb6f7-kube-api-access-nc669\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.525152 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-utilities\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.525246 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.02520461 +0000 UTC m=+159.999699430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.627324 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc669\" (UniqueName: \"kubernetes.io/projected/f75a50cf-7dae-420b-a00d-671ede3cb6f7-kube-api-access-nc669\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.627390 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-utilities\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.627475 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-catalog-content\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.627512 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.627925 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.127908619 +0000 UTC m=+160.102403439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.628240 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-utilities\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.628344 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-catalog-content\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.687850 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc669\" (UniqueName: \"kubernetes.io/projected/f75a50cf-7dae-420b-a00d-671ede3cb6f7-kube-api-access-nc669\") pod \"certified-operators-nrhl6\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.689868 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sbf5n"] Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.690966 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.719264 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sbf5n"] Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.729022 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.729233 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.229205583 +0000 UTC m=+160.203700413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.729343 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.729711 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.229704055 +0000 UTC m=+160.204198875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.771129 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.792257 4948 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-dkdgm container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.792658 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" podUID="a102d1fa-69a5-4826-ac40-e07ab4087558" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.809285 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-trh7g" event={"ID":"13b4eba0-42e7-448a-998c-d7aaeffc9218","Type":"ContainerStarted","Data":"8e92df1cce5fa37a86f5cbf1a1b9867cf0118f67442aabc21aee1d5d906263df"} Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.810149 4948 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-sjv8t container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.810214 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" podUID="95a270e5-7a08-4238-a65d-96986f2d2c39" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.814122 4948 patch_prober.go:28] interesting pod/downloads-7954f5f757-grnfc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.814190 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-grnfc" podUID="bc13d9fe-fd58-4d3b-9278-4f3e59da5976" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.832794 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.833052 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-utilities\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.833123 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqxvg\" (UniqueName: \"kubernetes.io/projected/000032f5-8df9-4879-8a8c-7ca06deb0862-kube-api-access-dqxvg\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.833155 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-catalog-content\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.833217 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.333185373 +0000 UTC m=+160.307680373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.833276 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.833757 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.333748297 +0000 UTC m=+160.308243127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.899145 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dkdgm" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.935058 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:30 crc kubenswrapper[4948]: E0220 08:08:30.935685 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.435645146 +0000 UTC m=+160.410139966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.937946 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-utilities\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.938151 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqxvg\" (UniqueName: \"kubernetes.io/projected/000032f5-8df9-4879-8a8c-7ca06deb0862-kube-api-access-dqxvg\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.938248 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-catalog-content\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.940081 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-catalog-content\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.941664 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-utilities\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:30 crc kubenswrapper[4948]: I0220 08:08:30.956304 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.041066 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.041525 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqxvg\" (UniqueName: \"kubernetes.io/projected/000032f5-8df9-4879-8a8c-7ca06deb0862-kube-api-access-dqxvg\") pod \"certified-operators-sbf5n\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.041838 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.54182311 +0000 UTC m=+160.516317930 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.083223 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n8d7m"] Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.087013 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.090959 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.091693 4948 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.121895 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n8d7m"] Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.144732 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.145303 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.645263637 +0000 UTC m=+160.619758467 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.145785 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-utilities\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.145961 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.146141 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxq9k\" (UniqueName: \"kubernetes.io/projected/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-kube-api-access-kxq9k\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.146235 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-catalog-content\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.146744 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.646725242 +0000 UTC m=+160.621220272 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.248680 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.249266 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxq9k\" (UniqueName: \"kubernetes.io/projected/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-kube-api-access-kxq9k\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.249307 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-catalog-content\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.249340 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-utilities\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.249816 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-utilities\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.249893 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.749876032 +0000 UTC m=+160.724370852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.250374 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-catalog-content\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.264229 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tvwkh"] Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.268100 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.294026 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tvwkh"] Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.297872 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxq9k\" (UniqueName: \"kubernetes.io/projected/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-kube-api-access-kxq9k\") pod \"community-operators-n8d7m\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.315336 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.350475 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-catalog-content\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.350515 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-utilities\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.350551 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.350592 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px4gp\" (UniqueName: \"kubernetes.io/projected/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-kube-api-access-px4gp\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.351613 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.850963601 +0000 UTC m=+160.825458421 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.423601 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.452598 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.452981 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-catalog-content\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.453018 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-utilities\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.453072 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px4gp\" (UniqueName: \"kubernetes.io/projected/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-kube-api-access-px4gp\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.453540 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:31.953522776 +0000 UTC m=+160.928017606 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.454342 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-utilities\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.454388 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-catalog-content\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.473452 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:31 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:31 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:31 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.473534 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.481013 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nrhl6"] Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.494886 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px4gp\" (UniqueName: \"kubernetes.io/projected/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-kube-api-access-px4gp\") pod \"community-operators-tvwkh\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.554901 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.557501 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:32.057477515 +0000 UTC m=+161.031972335 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.630335 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.664728 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.665331 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:32.165308919 +0000 UTC m=+161.139803739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.735399 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sbf5n"] Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.769825 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.770224 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:32.270202901 +0000 UTC m=+161.244697721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.814408 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n8d7m"] Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.853458 4948 generic.go:334] "Generic (PLEG): container finished" podID="f77016a6-cd35-49df-b6b5-65b4858b41c9" containerID="0520ff9e3a8f12a5228a352b59373978c12d5f10bb51923b12eab15acf564086" exitCode=0 Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.854780 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" event={"ID":"f77016a6-cd35-49df-b6b5-65b4858b41c9","Type":"ContainerDied","Data":"0520ff9e3a8f12a5228a352b59373978c12d5f10bb51923b12eab15acf564086"} Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.868121 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-trh7g" event={"ID":"13b4eba0-42e7-448a-998c-d7aaeffc9218","Type":"ContainerStarted","Data":"1f020d0f8752ef48a55c7d3b2aa1b9f3542e54bfa2d906bdd61229fe0bac45d4"} Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.871803 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.872152 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 08:08:32.372135161 +0000 UTC m=+161.346629981 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.875223 4948 generic.go:334] "Generic (PLEG): container finished" podID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerID="df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e" exitCode=0 Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.876785 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrhl6" event={"ID":"f75a50cf-7dae-420b-a00d-671ede3cb6f7","Type":"ContainerDied","Data":"df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e"} Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.876838 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrhl6" event={"ID":"f75a50cf-7dae-420b-a00d-671ede3cb6f7","Type":"ContainerStarted","Data":"d0015edf4c699a29515d340ae2c01786195a0e9d51c4459eae81be82178c452a"} Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.894025 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:08:31 crc kubenswrapper[4948]: I0220 08:08:31.997201 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:31 crc kubenswrapper[4948]: E0220 08:08:31.998054 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 08:08:32.498036159 +0000 UTC m=+161.472530979 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7wm8c" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.032312 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-trh7g" podStartSLOduration=11.032288829 podStartE2EDuration="11.032288829s" podCreationTimestamp="2026-02-20 08:08:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:31.999333431 +0000 UTC m=+160.973828251" watchObservedRunningTime="2026-02-20 08:08:32.032288829 +0000 UTC m=+161.006783649" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.081125 4948 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-20T08:08:31.091711343Z","Handler":null,"Name":""} Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.091538 4948 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.091598 4948 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.107738 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.115355 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.209583 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.318381 4948 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.318451 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.359334 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7wm8c\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.444186 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tvwkh"] Feb 20 08:08:32 crc kubenswrapper[4948]: W0220 08:08:32.449690 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6d696cd_f5cf_47e0_af8d_4d20e93bd80d.slice/crio-46adc95f753dc88a78d1ef713a2cbcbf1789c9d4ad7e29e46f1a4a26d4497aeb WatchSource:0}: Error finding container 46adc95f753dc88a78d1ef713a2cbcbf1789c9d4ad7e29e46f1a4a26d4497aeb: Status 404 returned error can't find the container with id 46adc95f753dc88a78d1ef713a2cbcbf1789c9d4ad7e29e46f1a4a26d4497aeb Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.462820 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:32 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:32 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:32 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.462893 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.594956 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.761658 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.855551 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44l7j" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.861715 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7wm8c"] Feb 20 08:08:32 crc kubenswrapper[4948]: W0220 08:08:32.876422 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57c455f7_aa9c_405c_bee3_89726b84f5db.slice/crio-f90694ba31dc9edaf32c68d5b113799c7c3cf39817bac7a06e98cc5e157c4d16 WatchSource:0}: Error finding container f90694ba31dc9edaf32c68d5b113799c7c3cf39817bac7a06e98cc5e157c4d16: Status 404 returned error can't find the container with id f90694ba31dc9edaf32c68d5b113799c7c3cf39817bac7a06e98cc5e157c4d16 Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.884240 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvwkh" event={"ID":"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d","Type":"ContainerStarted","Data":"46adc95f753dc88a78d1ef713a2cbcbf1789c9d4ad7e29e46f1a4a26d4497aeb"} Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.886338 4948 generic.go:334] "Generic (PLEG): container finished" podID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerID="b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93" exitCode=0 Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.886787 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sbf5n" event={"ID":"000032f5-8df9-4879-8a8c-7ca06deb0862","Type":"ContainerDied","Data":"b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93"} Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.886831 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sbf5n" event={"ID":"000032f5-8df9-4879-8a8c-7ca06deb0862","Type":"ContainerStarted","Data":"6682a9f05af301b707c145d9ce47468c3451f6e6fc6b07c7b78c99356b765594"} Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.897249 4948 generic.go:334] "Generic (PLEG): container finished" podID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerID="8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f" exitCode=0 Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.898243 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8d7m" event={"ID":"00f4b4c1-af36-4ccd-a3d1-dd6df1186338","Type":"ContainerDied","Data":"8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f"} Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.898359 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8d7m" event={"ID":"00f4b4c1-af36-4ccd-a3d1-dd6df1186338","Type":"ContainerStarted","Data":"33b053c8152742a922c7d8709831de18b8993625ae7594197966e98f173b1174"} Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.899246 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.900145 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.908602 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.909588 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 20 08:08:32 crc kubenswrapper[4948]: I0220 08:08:32.909770 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.024936 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b189bd1e-6355-433c-972d-2e27f98fe153-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b189bd1e-6355-433c-972d-2e27f98fe153\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.025014 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b189bd1e-6355-433c-972d-2e27f98fe153-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b189bd1e-6355-433c-972d-2e27f98fe153\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.090181 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c9z42"] Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.095167 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.097898 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.129299 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b189bd1e-6355-433c-972d-2e27f98fe153-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b189bd1e-6355-433c-972d-2e27f98fe153\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.129236 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b189bd1e-6355-433c-972d-2e27f98fe153-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b189bd1e-6355-433c-972d-2e27f98fe153\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.130672 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b189bd1e-6355-433c-972d-2e27f98fe153-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b189bd1e-6355-433c-972d-2e27f98fe153\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.149936 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9z42"] Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.171733 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b189bd1e-6355-433c-972d-2e27f98fe153-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b189bd1e-6355-433c-972d-2e27f98fe153\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.173457 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.181462 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.181488 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.201182 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.233652 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-utilities\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.233714 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqtvv\" (UniqueName: \"kubernetes.io/projected/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-kube-api-access-cqtvv\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.233770 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-catalog-content\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.324656 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.335272 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f77016a6-cd35-49df-b6b5-65b4858b41c9-secret-volume\") pod \"f77016a6-cd35-49df-b6b5-65b4858b41c9\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.335360 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ql4r\" (UniqueName: \"kubernetes.io/projected/f77016a6-cd35-49df-b6b5-65b4858b41c9-kube-api-access-2ql4r\") pod \"f77016a6-cd35-49df-b6b5-65b4858b41c9\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.335417 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f77016a6-cd35-49df-b6b5-65b4858b41c9-config-volume\") pod \"f77016a6-cd35-49df-b6b5-65b4858b41c9\" (UID: \"f77016a6-cd35-49df-b6b5-65b4858b41c9\") " Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.335686 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqtvv\" (UniqueName: \"kubernetes.io/projected/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-kube-api-access-cqtvv\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.335805 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-catalog-content\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.335958 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-utilities\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.336698 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-utilities\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.338718 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-catalog-content\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.341254 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f77016a6-cd35-49df-b6b5-65b4858b41c9-config-volume" (OuterVolumeSpecName: "config-volume") pod "f77016a6-cd35-49df-b6b5-65b4858b41c9" (UID: "f77016a6-cd35-49df-b6b5-65b4858b41c9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.341746 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f77016a6-cd35-49df-b6b5-65b4858b41c9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f77016a6-cd35-49df-b6b5-65b4858b41c9" (UID: "f77016a6-cd35-49df-b6b5-65b4858b41c9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.342745 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f77016a6-cd35-49df-b6b5-65b4858b41c9-kube-api-access-2ql4r" (OuterVolumeSpecName: "kube-api-access-2ql4r") pod "f77016a6-cd35-49df-b6b5-65b4858b41c9" (UID: "f77016a6-cd35-49df-b6b5-65b4858b41c9"). InnerVolumeSpecName "kube-api-access-2ql4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.357224 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqtvv\" (UniqueName: \"kubernetes.io/projected/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-kube-api-access-cqtvv\") pod \"redhat-marketplace-c9z42\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.422109 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.436930 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ql4r\" (UniqueName: \"kubernetes.io/projected/f77016a6-cd35-49df-b6b5-65b4858b41c9-kube-api-access-2ql4r\") on node \"crc\" DevicePath \"\"" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.436955 4948 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f77016a6-cd35-49df-b6b5-65b4858b41c9-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.436978 4948 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f77016a6-cd35-49df-b6b5-65b4858b41c9-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.467883 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:33 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:33 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:33 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.467948 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.469130 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-67ptx"] Feb 20 08:08:33 crc kubenswrapper[4948]: E0220 08:08:33.469919 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f77016a6-cd35-49df-b6b5-65b4858b41c9" containerName="collect-profiles" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.469939 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f77016a6-cd35-49df-b6b5-65b4858b41c9" containerName="collect-profiles" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.470042 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f77016a6-cd35-49df-b6b5-65b4858b41c9" containerName="collect-profiles" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.472247 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.479262 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-67ptx"] Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.579247 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 20 08:08:33 crc kubenswrapper[4948]: W0220 08:08:33.592343 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podb189bd1e_6355_433c_972d_2e27f98fe153.slice/crio-df6078b5b4a4366ce4b18c7e57e97fa910c6f7c0860321f7de0754fed2628438 WatchSource:0}: Error finding container df6078b5b4a4366ce4b18c7e57e97fa910c6f7c0860321f7de0754fed2628438: Status 404 returned error can't find the container with id df6078b5b4a4366ce4b18c7e57e97fa910c6f7c0860321f7de0754fed2628438 Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.639874 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-utilities\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.639918 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-catalog-content\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.640136 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx26r\" (UniqueName: \"kubernetes.io/projected/b8150552-4914-4e92-b041-56600ed6d2f5-kube-api-access-cx26r\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.664909 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9z42"] Feb 20 08:08:33 crc kubenswrapper[4948]: W0220 08:08:33.677193 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc53ddd33_0f9f_4794_b346_7d48a6c09c9b.slice/crio-9a7c5f1538f439aba0332cfbd20f5f653d7b240472d961eff61974598f86365b WatchSource:0}: Error finding container 9a7c5f1538f439aba0332cfbd20f5f653d7b240472d961eff61974598f86365b: Status 404 returned error can't find the container with id 9a7c5f1538f439aba0332cfbd20f5f653d7b240472d961eff61974598f86365b Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.733557 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.741466 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-utilities\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.741516 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-catalog-content\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.741630 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx26r\" (UniqueName: \"kubernetes.io/projected/b8150552-4914-4e92-b041-56600ed6d2f5-kube-api-access-cx26r\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.742567 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-utilities\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.742612 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-catalog-content\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.765782 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx26r\" (UniqueName: \"kubernetes.io/projected/b8150552-4914-4e92-b041-56600ed6d2f5-kube-api-access-cx26r\") pod \"redhat-marketplace-67ptx\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.777842 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.777903 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.785853 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.794484 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.818243 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.818311 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.820253 4948 patch_prober.go:28] interesting pod/console-f9d7485db-k8zcr container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.820360 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-k8zcr" podUID="f81366e5-fbdb-4a0b-a91f-f603d9b8570d" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.858606 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-frznz"] Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.859893 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.861894 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.878192 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-frznz"] Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.939935 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9z42" event={"ID":"c53ddd33-0f9f-4794-b346-7d48a6c09c9b","Type":"ContainerStarted","Data":"1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe"} Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.940063 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9z42" event={"ID":"c53ddd33-0f9f-4794-b346-7d48a6c09c9b","Type":"ContainerStarted","Data":"9a7c5f1538f439aba0332cfbd20f5f653d7b240472d961eff61974598f86365b"} Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.944480 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.944474 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk" event={"ID":"f77016a6-cd35-49df-b6b5-65b4858b41c9","Type":"ContainerDied","Data":"92efa091cdddd802dc78aa2d9bf128d9d0f0ac869c640718f1c1a8d1aead053f"} Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.944892 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92efa091cdddd802dc78aa2d9bf128d9d0f0ac869c640718f1c1a8d1aead053f" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.945440 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-utilities\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.945626 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-catalog-content\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.945789 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqs4x\" (UniqueName: \"kubernetes.io/projected/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-kube-api-access-hqs4x\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.947686 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" event={"ID":"57c455f7-aa9c-405c-bee3-89726b84f5db","Type":"ContainerStarted","Data":"20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120"} Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.947742 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" event={"ID":"57c455f7-aa9c-405c-bee3-89726b84f5db","Type":"ContainerStarted","Data":"f90694ba31dc9edaf32c68d5b113799c7c3cf39817bac7a06e98cc5e157c4d16"} Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.948106 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.950361 4948 generic.go:334] "Generic (PLEG): container finished" podID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerID="5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1" exitCode=0 Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.950442 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvwkh" event={"ID":"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d","Type":"ContainerDied","Data":"5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1"} Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.958607 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b189bd1e-6355-433c-972d-2e27f98fe153","Type":"ContainerStarted","Data":"df6078b5b4a4366ce4b18c7e57e97fa910c6f7c0860321f7de0754fed2628438"} Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.966267 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gf42w" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.968846 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-jfst2" Feb 20 08:08:33 crc kubenswrapper[4948]: I0220 08:08:33.991654 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" podStartSLOduration=139.991593848 podStartE2EDuration="2m19.991593848s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:33.987680922 +0000 UTC m=+162.962175742" watchObservedRunningTime="2026-02-20 08:08:33.991593848 +0000 UTC m=+162.966088668" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.049639 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqs4x\" (UniqueName: \"kubernetes.io/projected/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-kube-api-access-hqs4x\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.049734 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-utilities\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.049855 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-catalog-content\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.055556 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-catalog-content\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.058045 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-utilities\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.111451 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-snc6x"] Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.112780 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.114904 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqs4x\" (UniqueName: \"kubernetes.io/projected/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-kube-api-access-hqs4x\") pod \"redhat-operators-frznz\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.125033 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-snc6x"] Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.169031 4948 patch_prober.go:28] interesting pod/downloads-7954f5f757-grnfc container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.169400 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-grnfc" podUID="bc13d9fe-fd58-4d3b-9278-4f3e59da5976" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.180203 4948 patch_prober.go:28] interesting pod/downloads-7954f5f757-grnfc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.180279 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-grnfc" podUID="bc13d9fe-fd58-4d3b-9278-4f3e59da5976" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.215395 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.253109 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-catalog-content\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.253181 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc4n5\" (UniqueName: \"kubernetes.io/projected/b5ec9802-aa9c-4518-b99a-1673ab75ec50-kube-api-access-gc4n5\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.253266 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-utilities\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.351365 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-67ptx"] Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.361965 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-utilities\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.362094 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-catalog-content\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.362134 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc4n5\" (UniqueName: \"kubernetes.io/projected/b5ec9802-aa9c-4518-b99a-1673ab75ec50-kube-api-access-gc4n5\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.362449 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-utilities\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.362729 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-catalog-content\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.430775 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc4n5\" (UniqueName: \"kubernetes.io/projected/b5ec9802-aa9c-4518-b99a-1673ab75ec50-kube-api-access-gc4n5\") pod \"redhat-operators-snc6x\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.465632 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.466067 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.471438 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:34 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:34 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:34 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.471489 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.625395 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-frznz"] Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.630915 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:08:34 crc kubenswrapper[4948]: I0220 08:08:34.875555 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-snc6x"] Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.011070 4948 generic.go:334] "Generic (PLEG): container finished" podID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerID="6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7" exitCode=0 Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.011275 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frznz" event={"ID":"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9","Type":"ContainerDied","Data":"6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7"} Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.011537 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frznz" event={"ID":"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9","Type":"ContainerStarted","Data":"bc370df4b0b552966157b6ac4af296601fb860ed2508da9c4291e9924ec1e9cb"} Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.022917 4948 generic.go:334] "Generic (PLEG): container finished" podID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerID="1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe" exitCode=0 Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.023000 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9z42" event={"ID":"c53ddd33-0f9f-4794-b346-7d48a6c09c9b","Type":"ContainerDied","Data":"1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe"} Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.036600 4948 generic.go:334] "Generic (PLEG): container finished" podID="b8150552-4914-4e92-b041-56600ed6d2f5" containerID="24232a1b4e66920a8c674176ee38da24ffd8d17de7ae8a51e030daddaaeb552d" exitCode=0 Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.036709 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67ptx" event={"ID":"b8150552-4914-4e92-b041-56600ed6d2f5","Type":"ContainerDied","Data":"24232a1b4e66920a8c674176ee38da24ffd8d17de7ae8a51e030daddaaeb552d"} Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.037549 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67ptx" event={"ID":"b8150552-4914-4e92-b041-56600ed6d2f5","Type":"ContainerStarted","Data":"b740ba77241ab7ac59567dc96b03935ff231d4a2d8855de4a6ed59ddebaef76b"} Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.048551 4948 generic.go:334] "Generic (PLEG): container finished" podID="b189bd1e-6355-433c-972d-2e27f98fe153" containerID="e48b6bdd3e78b89a1786debf78466f80dde9962619930bae2b9cae797f181fd2" exitCode=0 Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.050675 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b189bd1e-6355-433c-972d-2e27f98fe153","Type":"ContainerDied","Data":"e48b6bdd3e78b89a1786debf78466f80dde9962619930bae2b9cae797f181fd2"} Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.219083 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.219939 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.224748 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.224748 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.226090 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.293136 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.293269 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.395367 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.395457 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.395855 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.420534 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.470340 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:35 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:35 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:35 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.470409 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.552308 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:35 crc kubenswrapper[4948]: I0220 08:08:35.916905 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 20 08:08:35 crc kubenswrapper[4948]: W0220 08:08:35.969608 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podc2d35955_eb5e_4d11_8e5a_5ed0716a5bc0.slice/crio-b80e153205aa2bfb003f43118045cec88da3439901cf54d41fc4f1ecf766667b WatchSource:0}: Error finding container b80e153205aa2bfb003f43118045cec88da3439901cf54d41fc4f1ecf766667b: Status 404 returned error can't find the container with id b80e153205aa2bfb003f43118045cec88da3439901cf54d41fc4f1ecf766667b Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.056921 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0","Type":"ContainerStarted","Data":"b80e153205aa2bfb003f43118045cec88da3439901cf54d41fc4f1ecf766667b"} Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.059280 4948 generic.go:334] "Generic (PLEG): container finished" podID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerID="5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320" exitCode=0 Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.059433 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snc6x" event={"ID":"b5ec9802-aa9c-4518-b99a-1673ab75ec50","Type":"ContainerDied","Data":"5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320"} Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.059527 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snc6x" event={"ID":"b5ec9802-aa9c-4518-b99a-1673ab75ec50","Type":"ContainerStarted","Data":"40f046d646c178228156ce2281e80f2e2be9092681446e798f497080898603ae"} Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.457846 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.464839 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:36 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:36 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:36 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.464904 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.522413 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b189bd1e-6355-433c-972d-2e27f98fe153-kube-api-access\") pod \"b189bd1e-6355-433c-972d-2e27f98fe153\" (UID: \"b189bd1e-6355-433c-972d-2e27f98fe153\") " Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.522510 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b189bd1e-6355-433c-972d-2e27f98fe153-kubelet-dir\") pod \"b189bd1e-6355-433c-972d-2e27f98fe153\" (UID: \"b189bd1e-6355-433c-972d-2e27f98fe153\") " Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.522792 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b189bd1e-6355-433c-972d-2e27f98fe153-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b189bd1e-6355-433c-972d-2e27f98fe153" (UID: "b189bd1e-6355-433c-972d-2e27f98fe153"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.541478 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b189bd1e-6355-433c-972d-2e27f98fe153-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b189bd1e-6355-433c-972d-2e27f98fe153" (UID: "b189bd1e-6355-433c-972d-2e27f98fe153"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.624383 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b189bd1e-6355-433c-972d-2e27f98fe153-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.624439 4948 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b189bd1e-6355-433c-972d-2e27f98fe153-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.725636 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.752756 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd-metrics-certs\") pod \"network-metrics-daemon-xd86g\" (UID: \"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd\") " pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:36 crc kubenswrapper[4948]: I0220 08:08:36.868170 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xd86g" Feb 20 08:08:37 crc kubenswrapper[4948]: I0220 08:08:37.135666 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0","Type":"ContainerStarted","Data":"e57316aee001d124820060d888331a36dba4f8f0e8e53102f7df766f39d91d8a"} Feb 20 08:08:37 crc kubenswrapper[4948]: I0220 08:08:37.149781 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b189bd1e-6355-433c-972d-2e27f98fe153","Type":"ContainerDied","Data":"df6078b5b4a4366ce4b18c7e57e97fa910c6f7c0860321f7de0754fed2628438"} Feb 20 08:08:37 crc kubenswrapper[4948]: I0220 08:08:37.149836 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df6078b5b4a4366ce4b18c7e57e97fa910c6f7c0860321f7de0754fed2628438" Feb 20 08:08:37 crc kubenswrapper[4948]: I0220 08:08:37.149875 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 08:08:37 crc kubenswrapper[4948]: I0220 08:08:37.158610 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.158588624 podStartE2EDuration="2.158588624s" podCreationTimestamp="2026-02-20 08:08:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:37.15391089 +0000 UTC m=+166.128405710" watchObservedRunningTime="2026-02-20 08:08:37.158588624 +0000 UTC m=+166.133083444" Feb 20 08:08:37 crc kubenswrapper[4948]: I0220 08:08:37.252680 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xd86g"] Feb 20 08:08:37 crc kubenswrapper[4948]: I0220 08:08:37.466400 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:37 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:37 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:37 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:37 crc kubenswrapper[4948]: I0220 08:08:37.466502 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:38 crc kubenswrapper[4948]: I0220 08:08:38.025370 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:08:38 crc kubenswrapper[4948]: I0220 08:08:38.025687 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:08:38 crc kubenswrapper[4948]: I0220 08:08:38.174110 4948 generic.go:334] "Generic (PLEG): container finished" podID="c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0" containerID="e57316aee001d124820060d888331a36dba4f8f0e8e53102f7df766f39d91d8a" exitCode=0 Feb 20 08:08:38 crc kubenswrapper[4948]: I0220 08:08:38.174174 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0","Type":"ContainerDied","Data":"e57316aee001d124820060d888331a36dba4f8f0e8e53102f7df766f39d91d8a"} Feb 20 08:08:38 crc kubenswrapper[4948]: I0220 08:08:38.203381 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xd86g" event={"ID":"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd","Type":"ContainerStarted","Data":"b1f89ee38d3a256c46d4126a49d66b2ee3a928d4035b624bbc59430d1a8c8dff"} Feb 20 08:08:38 crc kubenswrapper[4948]: I0220 08:08:38.461543 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:38 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:38 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:38 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:38 crc kubenswrapper[4948]: I0220 08:08:38.461620 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.224524 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xd86g" event={"ID":"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd","Type":"ContainerStarted","Data":"027950bf79b0645503bcffbc775f0c1cffa721ad8382c116eab7d302c5f8f2a2"} Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.225017 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xd86g" event={"ID":"62e5e340-02b4-4d4e-a5a1-bc3b917e5bdd","Type":"ContainerStarted","Data":"6bbc067891d2ea7f9b69d2c5053c14eecb4759d3497f466543ffce6b10dd81e9"} Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.263680 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-xd86g" podStartSLOduration=145.263651007 podStartE2EDuration="2m25.263651007s" podCreationTimestamp="2026-02-20 08:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:08:39.242065548 +0000 UTC m=+168.216560368" watchObservedRunningTime="2026-02-20 08:08:39.263651007 +0000 UTC m=+168.238145827" Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.466701 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:39 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:39 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:39 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.466766 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.633729 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.652886 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-9ls2l" Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.713317 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kube-api-access\") pod \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\" (UID: \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\") " Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.713492 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kubelet-dir\") pod \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\" (UID: \"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0\") " Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.713920 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0" (UID: "c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.715368 4948 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.739061 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0" (UID: "c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:08:39 crc kubenswrapper[4948]: I0220 08:08:39.816335 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 08:08:40 crc kubenswrapper[4948]: I0220 08:08:40.269611 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 08:08:40 crc kubenswrapper[4948]: I0220 08:08:40.270006 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0","Type":"ContainerDied","Data":"b80e153205aa2bfb003f43118045cec88da3439901cf54d41fc4f1ecf766667b"} Feb 20 08:08:40 crc kubenswrapper[4948]: I0220 08:08:40.270042 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b80e153205aa2bfb003f43118045cec88da3439901cf54d41fc4f1ecf766667b" Feb 20 08:08:40 crc kubenswrapper[4948]: I0220 08:08:40.462516 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:40 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:40 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:40 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:40 crc kubenswrapper[4948]: I0220 08:08:40.462600 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:41 crc kubenswrapper[4948]: I0220 08:08:41.463876 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:41 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:41 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:41 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:41 crc kubenswrapper[4948]: I0220 08:08:41.463953 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:42 crc kubenswrapper[4948]: I0220 08:08:42.462190 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:42 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:42 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:42 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:42 crc kubenswrapper[4948]: I0220 08:08:42.462272 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:43 crc kubenswrapper[4948]: I0220 08:08:43.462214 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:43 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:43 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:43 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:43 crc kubenswrapper[4948]: I0220 08:08:43.462567 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:43 crc kubenswrapper[4948]: I0220 08:08:43.817123 4948 patch_prober.go:28] interesting pod/console-f9d7485db-k8zcr container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Feb 20 08:08:43 crc kubenswrapper[4948]: I0220 08:08:43.817204 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-k8zcr" podUID="f81366e5-fbdb-4a0b-a91f-f603d9b8570d" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Feb 20 08:08:44 crc kubenswrapper[4948]: I0220 08:08:44.176380 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-grnfc" Feb 20 08:08:44 crc kubenswrapper[4948]: I0220 08:08:44.462103 4948 patch_prober.go:28] interesting pod/router-default-5444994796-sc5pm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 08:08:44 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Feb 20 08:08:44 crc kubenswrapper[4948]: [+]process-running ok Feb 20 08:08:44 crc kubenswrapper[4948]: healthz check failed Feb 20 08:08:44 crc kubenswrapper[4948]: I0220 08:08:44.462187 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sc5pm" podUID="4b3348f9-d9b9-4efb-9eb8-877f58b0d59c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 08:08:45 crc kubenswrapper[4948]: I0220 08:08:45.463167 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:45 crc kubenswrapper[4948]: I0220 08:08:45.470385 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-sc5pm" Feb 20 08:08:52 crc kubenswrapper[4948]: I0220 08:08:52.599644 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:08:53 crc kubenswrapper[4948]: I0220 08:08:53.822519 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:53 crc kubenswrapper[4948]: I0220 08:08:53.826827 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:08:57 crc kubenswrapper[4948]: E0220 08:08:57.182520 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 20 08:08:57 crc kubenswrapper[4948]: E0220 08:08:57.184354 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nc669,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-nrhl6_openshift-marketplace(f75a50cf-7dae-420b-a00d-671ede3cb6f7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 08:08:57 crc kubenswrapper[4948]: E0220 08:08:57.185664 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-nrhl6" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" Feb 20 08:08:58 crc kubenswrapper[4948]: E0220 08:08:58.716240 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-nrhl6" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" Feb 20 08:09:00 crc kubenswrapper[4948]: I0220 08:09:00.965278 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 08:09:04 crc kubenswrapper[4948]: I0220 08:09:04.275908 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-46sg8" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.627690 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.628342 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gc4n5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-snc6x_openshift-marketplace(b5ec9802-aa9c-4518-b99a-1673ab75ec50): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.629575 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-snc6x" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.690490 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.690719 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hqs4x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-frznz_openshift-marketplace(dc2fe1f3-084a-4e45-b64d-f385fb28d7e9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.691887 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-frznz" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.711226 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.711396 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cx26r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-67ptx_openshift-marketplace(b8150552-4914-4e92-b041-56600ed6d2f5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.712696 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-67ptx" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.713367 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.713529 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-px4gp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-tvwkh_openshift-marketplace(b6d696cd-f5cf-47e0-af8d-4d20e93bd80d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 08:09:05 crc kubenswrapper[4948]: E0220 08:09:05.714696 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-tvwkh" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" Feb 20 08:09:06 crc kubenswrapper[4948]: I0220 08:09:06.477987 4948 generic.go:334] "Generic (PLEG): container finished" podID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerID="5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7" exitCode=0 Feb 20 08:09:06 crc kubenswrapper[4948]: I0220 08:09:06.478016 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8d7m" event={"ID":"00f4b4c1-af36-4ccd-a3d1-dd6df1186338","Type":"ContainerDied","Data":"5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7"} Feb 20 08:09:06 crc kubenswrapper[4948]: I0220 08:09:06.483444 4948 generic.go:334] "Generic (PLEG): container finished" podID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerID="bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83" exitCode=0 Feb 20 08:09:06 crc kubenswrapper[4948]: I0220 08:09:06.483509 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9z42" event={"ID":"c53ddd33-0f9f-4794-b346-7d48a6c09c9b","Type":"ContainerDied","Data":"bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83"} Feb 20 08:09:06 crc kubenswrapper[4948]: I0220 08:09:06.486661 4948 generic.go:334] "Generic (PLEG): container finished" podID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerID="b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3" exitCode=0 Feb 20 08:09:06 crc kubenswrapper[4948]: I0220 08:09:06.487574 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sbf5n" event={"ID":"000032f5-8df9-4879-8a8c-7ca06deb0862","Type":"ContainerDied","Data":"b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3"} Feb 20 08:09:06 crc kubenswrapper[4948]: E0220 08:09:06.488913 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-frznz" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" Feb 20 08:09:06 crc kubenswrapper[4948]: E0220 08:09:06.491362 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-67ptx" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" Feb 20 08:09:06 crc kubenswrapper[4948]: E0220 08:09:06.491426 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-snc6x" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" Feb 20 08:09:06 crc kubenswrapper[4948]: E0220 08:09:06.493012 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-tvwkh" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" Feb 20 08:09:07 crc kubenswrapper[4948]: I0220 08:09:07.492808 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9z42" event={"ID":"c53ddd33-0f9f-4794-b346-7d48a6c09c9b","Type":"ContainerStarted","Data":"1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0"} Feb 20 08:09:07 crc kubenswrapper[4948]: I0220 08:09:07.515165 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c9z42" podStartSLOduration=2.43809356 podStartE2EDuration="34.515103266s" podCreationTimestamp="2026-02-20 08:08:33 +0000 UTC" firstStartedPulling="2026-02-20 08:08:35.076081143 +0000 UTC m=+164.050575963" lastFinishedPulling="2026-02-20 08:09:07.153090809 +0000 UTC m=+196.127585669" observedRunningTime="2026-02-20 08:09:07.512614805 +0000 UTC m=+196.487109635" watchObservedRunningTime="2026-02-20 08:09:07.515103266 +0000 UTC m=+196.489598086" Feb 20 08:09:08 crc kubenswrapper[4948]: I0220 08:09:08.025433 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:09:08 crc kubenswrapper[4948]: I0220 08:09:08.025913 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:09:08 crc kubenswrapper[4948]: I0220 08:09:08.500565 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sbf5n" event={"ID":"000032f5-8df9-4879-8a8c-7ca06deb0862","Type":"ContainerStarted","Data":"91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595"} Feb 20 08:09:08 crc kubenswrapper[4948]: I0220 08:09:08.505804 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8d7m" event={"ID":"00f4b4c1-af36-4ccd-a3d1-dd6df1186338","Type":"ContainerStarted","Data":"acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502"} Feb 20 08:09:08 crc kubenswrapper[4948]: I0220 08:09:08.522612 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sbf5n" podStartSLOduration=3.963252623 podStartE2EDuration="38.522594053s" podCreationTimestamp="2026-02-20 08:08:30 +0000 UTC" firstStartedPulling="2026-02-20 08:08:32.898081291 +0000 UTC m=+161.872576111" lastFinishedPulling="2026-02-20 08:09:07.457422721 +0000 UTC m=+196.431917541" observedRunningTime="2026-02-20 08:09:08.51961278 +0000 UTC m=+197.494107600" watchObservedRunningTime="2026-02-20 08:09:08.522594053 +0000 UTC m=+197.497088873" Feb 20 08:09:08 crc kubenswrapper[4948]: I0220 08:09:08.537279 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n8d7m" podStartSLOduration=3.028252157 podStartE2EDuration="37.537258553s" podCreationTimestamp="2026-02-20 08:08:31 +0000 UTC" firstStartedPulling="2026-02-20 08:08:32.909278346 +0000 UTC m=+161.883773166" lastFinishedPulling="2026-02-20 08:09:07.418284742 +0000 UTC m=+196.392779562" observedRunningTime="2026-02-20 08:09:08.53548235 +0000 UTC m=+197.509977170" watchObservedRunningTime="2026-02-20 08:09:08.537258553 +0000 UTC m=+197.511753373" Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.008776 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 20 08:09:10 crc kubenswrapper[4948]: E0220 08:09:10.011557 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b189bd1e-6355-433c-972d-2e27f98fe153" containerName="pruner" Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.011604 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b189bd1e-6355-433c-972d-2e27f98fe153" containerName="pruner" Feb 20 08:09:10 crc kubenswrapper[4948]: E0220 08:09:10.011637 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0" containerName="pruner" Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.011655 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0" containerName="pruner" Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.011856 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b189bd1e-6355-433c-972d-2e27f98fe153" containerName="pruner" Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.011895 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2d35955-eb5e-4d11-8e5a-5ed0716a5bc0" containerName="pruner" Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.012623 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.015953 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.017164 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 20 08:09:10 crc kubenswrapper[4948]: I0220 08:09:10.017807 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.204122 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.204205 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.306242 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.306415 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.306429 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.316743 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.316806 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.333032 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.424579 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.424642 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.434562 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.529044 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.530852 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.677082 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 20 08:09:11 crc kubenswrapper[4948]: I0220 08:09:11.995453 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sjnxm"] Feb 20 08:09:12 crc kubenswrapper[4948]: I0220 08:09:12.136043 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bb62d1ab-0c70-44c4-a350-ae1c566dbe89","Type":"ContainerStarted","Data":"b223cee57e1c4f9b3955bdbbeb144ea5ace687c6ba792bf62c3493ca138014b5"} Feb 20 08:09:12 crc kubenswrapper[4948]: I0220 08:09:12.136236 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bb62d1ab-0c70-44c4-a350-ae1c566dbe89","Type":"ContainerStarted","Data":"a9c38be594b02cd7ae1614b1e36587a85d7f59afa56ac8c0eed109f8018a0ec6"} Feb 20 08:09:12 crc kubenswrapper[4948]: I0220 08:09:12.164328 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=3.164307281 podStartE2EDuration="3.164307281s" podCreationTimestamp="2026-02-20 08:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:09:12.160409876 +0000 UTC m=+201.134904696" watchObservedRunningTime="2026-02-20 08:09:12.164307281 +0000 UTC m=+201.138802101" Feb 20 08:09:12 crc kubenswrapper[4948]: I0220 08:09:12.197921 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:09:12 crc kubenswrapper[4948]: I0220 08:09:12.202751 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:09:13 crc kubenswrapper[4948]: I0220 08:09:13.142125 4948 generic.go:334] "Generic (PLEG): container finished" podID="bb62d1ab-0c70-44c4-a350-ae1c566dbe89" containerID="b223cee57e1c4f9b3955bdbbeb144ea5ace687c6ba792bf62c3493ca138014b5" exitCode=0 Feb 20 08:09:13 crc kubenswrapper[4948]: I0220 08:09:13.142207 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bb62d1ab-0c70-44c4-a350-ae1c566dbe89","Type":"ContainerDied","Data":"b223cee57e1c4f9b3955bdbbeb144ea5ace687c6ba792bf62c3493ca138014b5"} Feb 20 08:09:13 crc kubenswrapper[4948]: I0220 08:09:13.144444 4948 generic.go:334] "Generic (PLEG): container finished" podID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerID="ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37" exitCode=0 Feb 20 08:09:13 crc kubenswrapper[4948]: I0220 08:09:13.144532 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrhl6" event={"ID":"f75a50cf-7dae-420b-a00d-671ede3cb6f7","Type":"ContainerDied","Data":"ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37"} Feb 20 08:09:13 crc kubenswrapper[4948]: I0220 08:09:13.422351 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:09:13 crc kubenswrapper[4948]: I0220 08:09:13.422526 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:09:13 crc kubenswrapper[4948]: I0220 08:09:13.467643 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.152140 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrhl6" event={"ID":"f75a50cf-7dae-420b-a00d-671ede3cb6f7","Type":"ContainerStarted","Data":"51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b"} Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.171473 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nrhl6" podStartSLOduration=2.258996389 podStartE2EDuration="44.171444284s" podCreationTimestamp="2026-02-20 08:08:30 +0000 UTC" firstStartedPulling="2026-02-20 08:08:31.893208258 +0000 UTC m=+160.867703078" lastFinishedPulling="2026-02-20 08:09:13.805656113 +0000 UTC m=+202.780150973" observedRunningTime="2026-02-20 08:09:14.169593548 +0000 UTC m=+203.144088378" watchObservedRunningTime="2026-02-20 08:09:14.171444284 +0000 UTC m=+203.145939124" Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.207729 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.398737 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.570593 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kubelet-dir\") pod \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\" (UID: \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\") " Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.570750 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bb62d1ab-0c70-44c4-a350-ae1c566dbe89" (UID: "bb62d1ab-0c70-44c4-a350-ae1c566dbe89"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.570780 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kube-api-access\") pod \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\" (UID: \"bb62d1ab-0c70-44c4-a350-ae1c566dbe89\") " Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.571847 4948 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.582137 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bb62d1ab-0c70-44c4-a350-ae1c566dbe89" (UID: "bb62d1ab-0c70-44c4-a350-ae1c566dbe89"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.672877 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb62d1ab-0c70-44c4-a350-ae1c566dbe89-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.713016 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sbf5n"] Feb 20 08:09:14 crc kubenswrapper[4948]: I0220 08:09:14.713241 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sbf5n" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerName="registry-server" containerID="cri-o://91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595" gracePeriod=2 Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.061426 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.158735 4948 generic.go:334] "Generic (PLEG): container finished" podID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerID="91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595" exitCode=0 Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.158802 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sbf5n" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.158808 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sbf5n" event={"ID":"000032f5-8df9-4879-8a8c-7ca06deb0862","Type":"ContainerDied","Data":"91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595"} Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.158842 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sbf5n" event={"ID":"000032f5-8df9-4879-8a8c-7ca06deb0862","Type":"ContainerDied","Data":"6682a9f05af301b707c145d9ce47468c3451f6e6fc6b07c7b78c99356b765594"} Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.158860 4948 scope.go:117] "RemoveContainer" containerID="91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.160494 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.160522 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bb62d1ab-0c70-44c4-a350-ae1c566dbe89","Type":"ContainerDied","Data":"a9c38be594b02cd7ae1614b1e36587a85d7f59afa56ac8c0eed109f8018a0ec6"} Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.160563 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9c38be594b02cd7ae1614b1e36587a85d7f59afa56ac8c0eed109f8018a0ec6" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.177828 4948 scope.go:117] "RemoveContainer" containerID="b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.179431 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqxvg\" (UniqueName: \"kubernetes.io/projected/000032f5-8df9-4879-8a8c-7ca06deb0862-kube-api-access-dqxvg\") pod \"000032f5-8df9-4879-8a8c-7ca06deb0862\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.179522 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-utilities\") pod \"000032f5-8df9-4879-8a8c-7ca06deb0862\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.179863 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-catalog-content\") pod \"000032f5-8df9-4879-8a8c-7ca06deb0862\" (UID: \"000032f5-8df9-4879-8a8c-7ca06deb0862\") " Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.180326 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-utilities" (OuterVolumeSpecName: "utilities") pod "000032f5-8df9-4879-8a8c-7ca06deb0862" (UID: "000032f5-8df9-4879-8a8c-7ca06deb0862"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.184176 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/000032f5-8df9-4879-8a8c-7ca06deb0862-kube-api-access-dqxvg" (OuterVolumeSpecName: "kube-api-access-dqxvg") pod "000032f5-8df9-4879-8a8c-7ca06deb0862" (UID: "000032f5-8df9-4879-8a8c-7ca06deb0862"). InnerVolumeSpecName "kube-api-access-dqxvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.224282 4948 scope.go:117] "RemoveContainer" containerID="b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.240957 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "000032f5-8df9-4879-8a8c-7ca06deb0862" (UID: "000032f5-8df9-4879-8a8c-7ca06deb0862"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.255638 4948 scope.go:117] "RemoveContainer" containerID="91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595" Feb 20 08:09:15 crc kubenswrapper[4948]: E0220 08:09:15.256135 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595\": container with ID starting with 91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595 not found: ID does not exist" containerID="91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.256183 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595"} err="failed to get container status \"91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595\": rpc error: code = NotFound desc = could not find container \"91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595\": container with ID starting with 91fe00d50fe88b88435954aa020922438c05af72280d83d2b0cbacc4ca3b0595 not found: ID does not exist" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.256242 4948 scope.go:117] "RemoveContainer" containerID="b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3" Feb 20 08:09:15 crc kubenswrapper[4948]: E0220 08:09:15.256750 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3\": container with ID starting with b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3 not found: ID does not exist" containerID="b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.256771 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3"} err="failed to get container status \"b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3\": rpc error: code = NotFound desc = could not find container \"b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3\": container with ID starting with b5372451792db1d5ea16e394e3175c7fb91527d66c015b609f73baf5c0f560d3 not found: ID does not exist" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.256785 4948 scope.go:117] "RemoveContainer" containerID="b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93" Feb 20 08:09:15 crc kubenswrapper[4948]: E0220 08:09:15.256989 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93\": container with ID starting with b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93 not found: ID does not exist" containerID="b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.257008 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93"} err="failed to get container status \"b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93\": rpc error: code = NotFound desc = could not find container \"b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93\": container with ID starting with b939d98be77246f833198ae82a9cce74eaf753fb0b14e37eb3663fc9e452cd93 not found: ID does not exist" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.281099 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqxvg\" (UniqueName: \"kubernetes.io/projected/000032f5-8df9-4879-8a8c-7ca06deb0862-kube-api-access-dqxvg\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.281136 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.281145 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000032f5-8df9-4879-8a8c-7ca06deb0862-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.487855 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sbf5n"] Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.492459 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sbf5n"] Feb 20 08:09:15 crc kubenswrapper[4948]: I0220 08:09:15.759597 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" path="/var/lib/kubelet/pods/000032f5-8df9-4879-8a8c-7ca06deb0862/volumes" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.002100 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 20 08:09:16 crc kubenswrapper[4948]: E0220 08:09:16.002392 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerName="extract-content" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.002418 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerName="extract-content" Feb 20 08:09:16 crc kubenswrapper[4948]: E0220 08:09:16.002434 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerName="extract-utilities" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.002445 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerName="extract-utilities" Feb 20 08:09:16 crc kubenswrapper[4948]: E0220 08:09:16.002459 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb62d1ab-0c70-44c4-a350-ae1c566dbe89" containerName="pruner" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.002471 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb62d1ab-0c70-44c4-a350-ae1c566dbe89" containerName="pruner" Feb 20 08:09:16 crc kubenswrapper[4948]: E0220 08:09:16.002489 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerName="registry-server" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.002499 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerName="registry-server" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.002641 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb62d1ab-0c70-44c4-a350-ae1c566dbe89" containerName="pruner" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.002658 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="000032f5-8df9-4879-8a8c-7ca06deb0862" containerName="registry-server" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.003145 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.005003 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.005150 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.013241 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.191461 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.191519 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-var-lock\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.191552 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kube-api-access\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.292853 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.292895 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.292965 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-var-lock\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.293054 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kube-api-access\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.293158 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-var-lock\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.312698 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kube-api-access\") pod \"installer-9-crc\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.323337 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:16 crc kubenswrapper[4948]: I0220 08:09:16.781292 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 20 08:09:16 crc kubenswrapper[4948]: W0220 08:09:16.790601 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda0fc48a9_8f2e_4a58_9ec2_0ab67f6cedea.slice/crio-501f50c7bcee3c3ff4d799d7aadad96a2760427edf1eb9577fa78b654115cf07 WatchSource:0}: Error finding container 501f50c7bcee3c3ff4d799d7aadad96a2760427edf1eb9577fa78b654115cf07: Status 404 returned error can't find the container with id 501f50c7bcee3c3ff4d799d7aadad96a2760427edf1eb9577fa78b654115cf07 Feb 20 08:09:17 crc kubenswrapper[4948]: I0220 08:09:17.173211 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea","Type":"ContainerStarted","Data":"501f50c7bcee3c3ff4d799d7aadad96a2760427edf1eb9577fa78b654115cf07"} Feb 20 08:09:18 crc kubenswrapper[4948]: I0220 08:09:18.179885 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea","Type":"ContainerStarted","Data":"308dd2b5632b47b59d9d195703329e9e330103427c0967b20ee97ba81766b1b3"} Feb 20 08:09:18 crc kubenswrapper[4948]: I0220 08:09:18.198821 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.198793508 podStartE2EDuration="3.198793508s" podCreationTimestamp="2026-02-20 08:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:09:18.197170358 +0000 UTC m=+207.171665218" watchObservedRunningTime="2026-02-20 08:09:18.198793508 +0000 UTC m=+207.173288368" Feb 20 08:09:20 crc kubenswrapper[4948]: I0220 08:09:20.772494 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:09:20 crc kubenswrapper[4948]: I0220 08:09:20.773606 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:09:20 crc kubenswrapper[4948]: I0220 08:09:20.813608 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:09:21 crc kubenswrapper[4948]: I0220 08:09:21.313391 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:09:27 crc kubenswrapper[4948]: I0220 08:09:27.315934 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snc6x" event={"ID":"b5ec9802-aa9c-4518-b99a-1673ab75ec50","Type":"ContainerStarted","Data":"e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e"} Feb 20 08:09:27 crc kubenswrapper[4948]: I0220 08:09:27.318366 4948 generic.go:334] "Generic (PLEG): container finished" podID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerID="adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893" exitCode=0 Feb 20 08:09:27 crc kubenswrapper[4948]: I0220 08:09:27.318416 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvwkh" event={"ID":"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d","Type":"ContainerDied","Data":"adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893"} Feb 20 08:09:28 crc kubenswrapper[4948]: I0220 08:09:28.326270 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frznz" event={"ID":"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9","Type":"ContainerStarted","Data":"27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00"} Feb 20 08:09:28 crc kubenswrapper[4948]: I0220 08:09:28.328875 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67ptx" event={"ID":"b8150552-4914-4e92-b041-56600ed6d2f5","Type":"ContainerStarted","Data":"e9c973e2d8ac9c5ce88b599645da506dbb8f6dbaabbec1d3e5fca44ff09a3d5c"} Feb 20 08:09:28 crc kubenswrapper[4948]: I0220 08:09:28.330782 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snc6x" event={"ID":"b5ec9802-aa9c-4518-b99a-1673ab75ec50","Type":"ContainerDied","Data":"e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e"} Feb 20 08:09:28 crc kubenswrapper[4948]: I0220 08:09:28.330681 4948 generic.go:334] "Generic (PLEG): container finished" podID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerID="e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e" exitCode=0 Feb 20 08:09:29 crc kubenswrapper[4948]: I0220 08:09:29.339519 4948 generic.go:334] "Generic (PLEG): container finished" podID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerID="27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00" exitCode=0 Feb 20 08:09:29 crc kubenswrapper[4948]: I0220 08:09:29.339619 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frznz" event={"ID":"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9","Type":"ContainerDied","Data":"27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00"} Feb 20 08:09:29 crc kubenswrapper[4948]: I0220 08:09:29.343571 4948 generic.go:334] "Generic (PLEG): container finished" podID="b8150552-4914-4e92-b041-56600ed6d2f5" containerID="e9c973e2d8ac9c5ce88b599645da506dbb8f6dbaabbec1d3e5fca44ff09a3d5c" exitCode=0 Feb 20 08:09:29 crc kubenswrapper[4948]: I0220 08:09:29.343657 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67ptx" event={"ID":"b8150552-4914-4e92-b041-56600ed6d2f5","Type":"ContainerDied","Data":"e9c973e2d8ac9c5ce88b599645da506dbb8f6dbaabbec1d3e5fca44ff09a3d5c"} Feb 20 08:09:29 crc kubenswrapper[4948]: I0220 08:09:29.346273 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snc6x" event={"ID":"b5ec9802-aa9c-4518-b99a-1673ab75ec50","Type":"ContainerStarted","Data":"59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e"} Feb 20 08:09:29 crc kubenswrapper[4948]: I0220 08:09:29.348992 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvwkh" event={"ID":"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d","Type":"ContainerStarted","Data":"3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d"} Feb 20 08:09:29 crc kubenswrapper[4948]: I0220 08:09:29.427344 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-snc6x" podStartSLOduration=2.684655272 podStartE2EDuration="55.42730509s" podCreationTimestamp="2026-02-20 08:08:34 +0000 UTC" firstStartedPulling="2026-02-20 08:08:36.067381954 +0000 UTC m=+165.041876774" lastFinishedPulling="2026-02-20 08:09:28.810031772 +0000 UTC m=+217.784526592" observedRunningTime="2026-02-20 08:09:29.398590686 +0000 UTC m=+218.373085526" watchObservedRunningTime="2026-02-20 08:09:29.42730509 +0000 UTC m=+218.401799930" Feb 20 08:09:29 crc kubenswrapper[4948]: I0220 08:09:29.451205 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tvwkh" podStartSLOduration=3.612786143 podStartE2EDuration="58.451179756s" podCreationTimestamp="2026-02-20 08:08:31 +0000 UTC" firstStartedPulling="2026-02-20 08:08:33.957213385 +0000 UTC m=+162.931708205" lastFinishedPulling="2026-02-20 08:09:28.795606998 +0000 UTC m=+217.770101818" observedRunningTime="2026-02-20 08:09:29.44893173 +0000 UTC m=+218.423426550" watchObservedRunningTime="2026-02-20 08:09:29.451179756 +0000 UTC m=+218.425674576" Feb 20 08:09:31 crc kubenswrapper[4948]: I0220 08:09:31.362526 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frznz" event={"ID":"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9","Type":"ContainerStarted","Data":"fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd"} Feb 20 08:09:31 crc kubenswrapper[4948]: I0220 08:09:31.365223 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67ptx" event={"ID":"b8150552-4914-4e92-b041-56600ed6d2f5","Type":"ContainerStarted","Data":"3ece74a70c863ce99b2f2f650f2d87de55dd27277341567aa5ed99e67f8a4dc8"} Feb 20 08:09:31 crc kubenswrapper[4948]: I0220 08:09:31.407114 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-frznz" podStartSLOduration=3.17650005 podStartE2EDuration="58.407089222s" podCreationTimestamp="2026-02-20 08:08:33 +0000 UTC" firstStartedPulling="2026-02-20 08:08:35.075987961 +0000 UTC m=+164.050482781" lastFinishedPulling="2026-02-20 08:09:30.306577123 +0000 UTC m=+219.281071953" observedRunningTime="2026-02-20 08:09:31.385935723 +0000 UTC m=+220.360430543" watchObservedRunningTime="2026-02-20 08:09:31.407089222 +0000 UTC m=+220.381584052" Feb 20 08:09:31 crc kubenswrapper[4948]: I0220 08:09:31.408878 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-67ptx" podStartSLOduration=2.779394732 podStartE2EDuration="58.408867796s" podCreationTimestamp="2026-02-20 08:08:33 +0000 UTC" firstStartedPulling="2026-02-20 08:08:35.0759414 +0000 UTC m=+164.050436220" lastFinishedPulling="2026-02-20 08:09:30.705414424 +0000 UTC m=+219.679909284" observedRunningTime="2026-02-20 08:09:31.404619171 +0000 UTC m=+220.379114001" watchObservedRunningTime="2026-02-20 08:09:31.408867796 +0000 UTC m=+220.383362636" Feb 20 08:09:31 crc kubenswrapper[4948]: I0220 08:09:31.631557 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:09:31 crc kubenswrapper[4948]: I0220 08:09:31.631632 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:09:31 crc kubenswrapper[4948]: I0220 08:09:31.688041 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:09:33 crc kubenswrapper[4948]: I0220 08:09:33.795831 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:09:33 crc kubenswrapper[4948]: I0220 08:09:33.796950 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:09:33 crc kubenswrapper[4948]: I0220 08:09:33.860217 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:09:34 crc kubenswrapper[4948]: I0220 08:09:34.216224 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:09:34 crc kubenswrapper[4948]: I0220 08:09:34.216278 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:09:34 crc kubenswrapper[4948]: I0220 08:09:34.467037 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:09:34 crc kubenswrapper[4948]: I0220 08:09:34.467355 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:09:35 crc kubenswrapper[4948]: I0220 08:09:35.257044 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-frznz" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="registry-server" probeResult="failure" output=< Feb 20 08:09:35 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 08:09:35 crc kubenswrapper[4948]: > Feb 20 08:09:35 crc kubenswrapper[4948]: I0220 08:09:35.505444 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-snc6x" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="registry-server" probeResult="failure" output=< Feb 20 08:09:35 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 08:09:35 crc kubenswrapper[4948]: > Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.030030 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" podUID="df73228e-df1f-4e9c-aafe-493bb743f98a" containerName="oauth-openshift" containerID="cri-o://f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30" gracePeriod=15 Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.380893 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.396466 4948 generic.go:334] "Generic (PLEG): container finished" podID="df73228e-df1f-4e9c-aafe-493bb743f98a" containerID="f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30" exitCode=0 Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.396510 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" event={"ID":"df73228e-df1f-4e9c-aafe-493bb743f98a","Type":"ContainerDied","Data":"f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30"} Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.396533 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" event={"ID":"df73228e-df1f-4e9c-aafe-493bb743f98a","Type":"ContainerDied","Data":"76dafdb5407bf70db2d0cf3c70d0d7427c429ce3ab8c2b2e1a871cb56382ee9a"} Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.396549 4948 scope.go:117] "RemoveContainer" containerID="f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.396669 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sjnxm" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.417701 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-d96b794dc-r7qn6"] Feb 20 08:09:37 crc kubenswrapper[4948]: E0220 08:09:37.418313 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df73228e-df1f-4e9c-aafe-493bb743f98a" containerName="oauth-openshift" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.418326 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="df73228e-df1f-4e9c-aafe-493bb743f98a" containerName="oauth-openshift" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.418452 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="df73228e-df1f-4e9c-aafe-493bb743f98a" containerName="oauth-openshift" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.418909 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.421388 4948 scope.go:117] "RemoveContainer" containerID="f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30" Feb 20 08:09:37 crc kubenswrapper[4948]: E0220 08:09:37.421847 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30\": container with ID starting with f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30 not found: ID does not exist" containerID="f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.421871 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30"} err="failed to get container status \"f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30\": rpc error: code = NotFound desc = could not find container \"f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30\": container with ID starting with f0e4fad62991d25dea391bf5b9e653af0d194dd88ba8a814b4a27d229ce2ef30 not found: ID does not exist" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.445480 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-ocp-branding-template\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446004 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-service-ca\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446144 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2l9z\" (UniqueName: \"kubernetes.io/projected/df73228e-df1f-4e9c-aafe-493bb743f98a-kube-api-access-q2l9z\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446430 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-service-ca\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446459 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-login\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446479 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5f01646f-0fb2-40ee-81aa-77f177add987-audit-dir\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446665 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446690 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446828 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-session\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.446859 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.447038 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-router-certs\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.447058 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czs2d\" (UniqueName: \"kubernetes.io/projected/5f01646f-0fb2-40ee-81aa-77f177add987-kube-api-access-czs2d\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.447078 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-audit-policies\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.447276 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.447300 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.447320 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-error\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.447448 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.448075 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.453456 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.454062 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df73228e-df1f-4e9c-aafe-493bb743f98a-kube-api-access-q2l9z" (OuterVolumeSpecName: "kube-api-access-q2l9z") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "kube-api-access-q2l9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.474543 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-d96b794dc-r7qn6"] Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548386 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-provider-selection\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548507 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-trusted-ca-bundle\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548579 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-cliconfig\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548624 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-session\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548656 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-error\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548678 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-login\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548715 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-idp-0-file-data\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548744 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-dir\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548770 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-policies\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548796 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-router-certs\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548816 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-serving-cert\") pod \"df73228e-df1f-4e9c-aafe-493bb743f98a\" (UID: \"df73228e-df1f-4e9c-aafe-493bb743f98a\") " Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.548958 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549007 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-router-certs\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549036 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czs2d\" (UniqueName: \"kubernetes.io/projected/5f01646f-0fb2-40ee-81aa-77f177add987-kube-api-access-czs2d\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549061 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-audit-policies\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549083 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549103 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549123 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-error\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549149 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549171 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-service-ca\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549200 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-login\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549219 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5f01646f-0fb2-40ee-81aa-77f177add987-audit-dir\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549241 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549261 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549288 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-session\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549322 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549334 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2l9z\" (UniqueName: \"kubernetes.io/projected/df73228e-df1f-4e9c-aafe-493bb743f98a-kube-api-access-q2l9z\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549345 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.549443 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.551134 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-audit-policies\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.551225 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5f01646f-0fb2-40ee-81aa-77f177add987-audit-dir\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.551360 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.551673 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.552204 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-service-ca\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.552752 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.552854 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.554132 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.557606 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.559191 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-router-certs\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.559292 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-session\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.559931 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.561162 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.570189 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.570296 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-login\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.570396 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.570625 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-error\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.570791 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.570961 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.571185 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.571535 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.571676 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "df73228e-df1f-4e9c-aafe-493bb743f98a" (UID: "df73228e-df1f-4e9c-aafe-493bb743f98a"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.574310 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5f01646f-0fb2-40ee-81aa-77f177add987-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.574617 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czs2d\" (UniqueName: \"kubernetes.io/projected/5f01646f-0fb2-40ee-81aa-77f177add987-kube-api-access-czs2d\") pod \"oauth-openshift-d96b794dc-r7qn6\" (UID: \"5f01646f-0fb2-40ee-81aa-77f177add987\") " pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651137 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651198 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651221 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651242 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651262 4948 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651280 4948 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651298 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651315 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651335 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651353 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.651371 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/df73228e-df1f-4e9c-aafe-493bb743f98a-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.733358 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.748407 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sjnxm"] Feb 20 08:09:37 crc kubenswrapper[4948]: I0220 08:09:37.750475 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sjnxm"] Feb 20 08:09:38 crc kubenswrapper[4948]: I0220 08:09:38.025672 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:09:38 crc kubenswrapper[4948]: I0220 08:09:38.025770 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:09:38 crc kubenswrapper[4948]: I0220 08:09:38.025839 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:09:38 crc kubenswrapper[4948]: I0220 08:09:38.026960 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:09:38 crc kubenswrapper[4948]: I0220 08:09:38.027103 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59" gracePeriod=600 Feb 20 08:09:38 crc kubenswrapper[4948]: I0220 08:09:38.257337 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-d96b794dc-r7qn6"] Feb 20 08:09:38 crc kubenswrapper[4948]: W0220 08:09:38.265512 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f01646f_0fb2_40ee_81aa_77f177add987.slice/crio-0b2fc2bc7a6727af6636f96647e0bcbf84785fef9ee71daef8751c86bc59125b WatchSource:0}: Error finding container 0b2fc2bc7a6727af6636f96647e0bcbf84785fef9ee71daef8751c86bc59125b: Status 404 returned error can't find the container with id 0b2fc2bc7a6727af6636f96647e0bcbf84785fef9ee71daef8751c86bc59125b Feb 20 08:09:38 crc kubenswrapper[4948]: I0220 08:09:38.408412 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" event={"ID":"5f01646f-0fb2-40ee-81aa-77f177add987","Type":"ContainerStarted","Data":"0b2fc2bc7a6727af6636f96647e0bcbf84785fef9ee71daef8751c86bc59125b"} Feb 20 08:09:39 crc kubenswrapper[4948]: I0220 08:09:39.420498 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59" exitCode=0 Feb 20 08:09:39 crc kubenswrapper[4948]: I0220 08:09:39.420598 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59"} Feb 20 08:09:39 crc kubenswrapper[4948]: I0220 08:09:39.421145 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"6791802a782456e3c45df951fa55097ce0ea3affd6a9878fc2981511046d1b09"} Feb 20 08:09:39 crc kubenswrapper[4948]: I0220 08:09:39.424378 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" event={"ID":"5f01646f-0fb2-40ee-81aa-77f177add987","Type":"ContainerStarted","Data":"8e95464b61e8b22fef4ef903732e5929f250c789b181c6c0c192edf0fbeb96ba"} Feb 20 08:09:39 crc kubenswrapper[4948]: I0220 08:09:39.427177 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:39 crc kubenswrapper[4948]: I0220 08:09:39.434101 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" Feb 20 08:09:39 crc kubenswrapper[4948]: I0220 08:09:39.729725 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df73228e-df1f-4e9c-aafe-493bb743f98a" path="/var/lib/kubelet/pods/df73228e-df1f-4e9c-aafe-493bb743f98a/volumes" Feb 20 08:09:41 crc kubenswrapper[4948]: I0220 08:09:41.710964 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:09:41 crc kubenswrapper[4948]: I0220 08:09:41.751385 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-d96b794dc-r7qn6" podStartSLOduration=29.751347097 podStartE2EDuration="29.751347097s" podCreationTimestamp="2026-02-20 08:09:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:09:39.482487568 +0000 UTC m=+228.456982428" watchObservedRunningTime="2026-02-20 08:09:41.751347097 +0000 UTC m=+230.725841967" Feb 20 08:09:41 crc kubenswrapper[4948]: I0220 08:09:41.774605 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tvwkh"] Feb 20 08:09:42 crc kubenswrapper[4948]: I0220 08:09:42.441583 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tvwkh" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerName="registry-server" containerID="cri-o://3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d" gracePeriod=2 Feb 20 08:09:42 crc kubenswrapper[4948]: I0220 08:09:42.877426 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.045477 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px4gp\" (UniqueName: \"kubernetes.io/projected/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-kube-api-access-px4gp\") pod \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.045608 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-catalog-content\") pod \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.045646 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-utilities\") pod \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\" (UID: \"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d\") " Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.047600 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-utilities" (OuterVolumeSpecName: "utilities") pod "b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" (UID: "b6d696cd-f5cf-47e0-af8d-4d20e93bd80d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.056118 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-kube-api-access-px4gp" (OuterVolumeSpecName: "kube-api-access-px4gp") pod "b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" (UID: "b6d696cd-f5cf-47e0-af8d-4d20e93bd80d"). InnerVolumeSpecName "kube-api-access-px4gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.114911 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" (UID: "b6d696cd-f5cf-47e0-af8d-4d20e93bd80d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.147152 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px4gp\" (UniqueName: \"kubernetes.io/projected/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-kube-api-access-px4gp\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.147199 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.147216 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.450513 4948 generic.go:334] "Generic (PLEG): container finished" podID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerID="3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d" exitCode=0 Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.451188 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvwkh" event={"ID":"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d","Type":"ContainerDied","Data":"3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d"} Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.451295 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvwkh" event={"ID":"b6d696cd-f5cf-47e0-af8d-4d20e93bd80d","Type":"ContainerDied","Data":"46adc95f753dc88a78d1ef713a2cbcbf1789c9d4ad7e29e46f1a4a26d4497aeb"} Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.451306 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvwkh" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.451333 4948 scope.go:117] "RemoveContainer" containerID="3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.488602 4948 scope.go:117] "RemoveContainer" containerID="adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.500445 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tvwkh"] Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.512836 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tvwkh"] Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.526251 4948 scope.go:117] "RemoveContainer" containerID="5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.548168 4948 scope.go:117] "RemoveContainer" containerID="3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d" Feb 20 08:09:43 crc kubenswrapper[4948]: E0220 08:09:43.548764 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d\": container with ID starting with 3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d not found: ID does not exist" containerID="3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.548827 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d"} err="failed to get container status \"3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d\": rpc error: code = NotFound desc = could not find container \"3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d\": container with ID starting with 3eb57d35e2ca5917ed63ba672507ebdc1b840d438dcdc8930f619e051d58336d not found: ID does not exist" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.548866 4948 scope.go:117] "RemoveContainer" containerID="adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893" Feb 20 08:09:43 crc kubenswrapper[4948]: E0220 08:09:43.549537 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893\": container with ID starting with adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893 not found: ID does not exist" containerID="adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.549603 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893"} err="failed to get container status \"adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893\": rpc error: code = NotFound desc = could not find container \"adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893\": container with ID starting with adec323384ee4b89adc829f6e8cec6c7df59b4806363aa0408583b7df2842893 not found: ID does not exist" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.549666 4948 scope.go:117] "RemoveContainer" containerID="5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1" Feb 20 08:09:43 crc kubenswrapper[4948]: E0220 08:09:43.550276 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1\": container with ID starting with 5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1 not found: ID does not exist" containerID="5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.550336 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1"} err="failed to get container status \"5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1\": rpc error: code = NotFound desc = could not find container \"5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1\": container with ID starting with 5e2b81703211ae6dafce6a9ca7e54224c7be8c36c969c7e7bf86ed1763e784b1 not found: ID does not exist" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.732112 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" path="/var/lib/kubelet/pods/b6d696cd-f5cf-47e0-af8d-4d20e93bd80d/volumes" Feb 20 08:09:43 crc kubenswrapper[4948]: I0220 08:09:43.856146 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:09:44 crc kubenswrapper[4948]: I0220 08:09:44.297908 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:09:44 crc kubenswrapper[4948]: I0220 08:09:44.342945 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:09:44 crc kubenswrapper[4948]: I0220 08:09:44.515799 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:09:44 crc kubenswrapper[4948]: I0220 08:09:44.586891 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.152795 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-67ptx"] Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.153597 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-67ptx" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" containerName="registry-server" containerID="cri-o://3ece74a70c863ce99b2f2f650f2d87de55dd27277341567aa5ed99e67f8a4dc8" gracePeriod=2 Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.479404 4948 generic.go:334] "Generic (PLEG): container finished" podID="b8150552-4914-4e92-b041-56600ed6d2f5" containerID="3ece74a70c863ce99b2f2f650f2d87de55dd27277341567aa5ed99e67f8a4dc8" exitCode=0 Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.479475 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67ptx" event={"ID":"b8150552-4914-4e92-b041-56600ed6d2f5","Type":"ContainerDied","Data":"3ece74a70c863ce99b2f2f650f2d87de55dd27277341567aa5ed99e67f8a4dc8"} Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.573639 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.697449 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cx26r\" (UniqueName: \"kubernetes.io/projected/b8150552-4914-4e92-b041-56600ed6d2f5-kube-api-access-cx26r\") pod \"b8150552-4914-4e92-b041-56600ed6d2f5\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.697530 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-utilities\") pod \"b8150552-4914-4e92-b041-56600ed6d2f5\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.697616 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-catalog-content\") pod \"b8150552-4914-4e92-b041-56600ed6d2f5\" (UID: \"b8150552-4914-4e92-b041-56600ed6d2f5\") " Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.698779 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-utilities" (OuterVolumeSpecName: "utilities") pod "b8150552-4914-4e92-b041-56600ed6d2f5" (UID: "b8150552-4914-4e92-b041-56600ed6d2f5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.704061 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8150552-4914-4e92-b041-56600ed6d2f5-kube-api-access-cx26r" (OuterVolumeSpecName: "kube-api-access-cx26r") pod "b8150552-4914-4e92-b041-56600ed6d2f5" (UID: "b8150552-4914-4e92-b041-56600ed6d2f5"). InnerVolumeSpecName "kube-api-access-cx26r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.734109 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b8150552-4914-4e92-b041-56600ed6d2f5" (UID: "b8150552-4914-4e92-b041-56600ed6d2f5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.751008 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-snc6x"] Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.751218 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-snc6x" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="registry-server" containerID="cri-o://59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e" gracePeriod=2 Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.799389 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.799690 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cx26r\" (UniqueName: \"kubernetes.io/projected/b8150552-4914-4e92-b041-56600ed6d2f5-kube-api-access-cx26r\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:46 crc kubenswrapper[4948]: I0220 08:09:46.799709 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8150552-4914-4e92-b041-56600ed6d2f5-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.076729 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.208281 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-utilities\") pod \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.208733 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-catalog-content\") pod \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.208910 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gc4n5\" (UniqueName: \"kubernetes.io/projected/b5ec9802-aa9c-4518-b99a-1673ab75ec50-kube-api-access-gc4n5\") pod \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\" (UID: \"b5ec9802-aa9c-4518-b99a-1673ab75ec50\") " Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.209297 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-utilities" (OuterVolumeSpecName: "utilities") pod "b5ec9802-aa9c-4518-b99a-1673ab75ec50" (UID: "b5ec9802-aa9c-4518-b99a-1673ab75ec50"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.214282 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5ec9802-aa9c-4518-b99a-1673ab75ec50-kube-api-access-gc4n5" (OuterVolumeSpecName: "kube-api-access-gc4n5") pod "b5ec9802-aa9c-4518-b99a-1673ab75ec50" (UID: "b5ec9802-aa9c-4518-b99a-1673ab75ec50"). InnerVolumeSpecName "kube-api-access-gc4n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.309852 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gc4n5\" (UniqueName: \"kubernetes.io/projected/b5ec9802-aa9c-4518-b99a-1673ab75ec50-kube-api-access-gc4n5\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.309890 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.340858 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5ec9802-aa9c-4518-b99a-1673ab75ec50" (UID: "b5ec9802-aa9c-4518-b99a-1673ab75ec50"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.411644 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5ec9802-aa9c-4518-b99a-1673ab75ec50-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.487085 4948 generic.go:334] "Generic (PLEG): container finished" podID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerID="59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e" exitCode=0 Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.487164 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snc6x" event={"ID":"b5ec9802-aa9c-4518-b99a-1673ab75ec50","Type":"ContainerDied","Data":"59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e"} Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.487199 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snc6x" event={"ID":"b5ec9802-aa9c-4518-b99a-1673ab75ec50","Type":"ContainerDied","Data":"40f046d646c178228156ce2281e80f2e2be9092681446e798f497080898603ae"} Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.487220 4948 scope.go:117] "RemoveContainer" containerID="59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.487247 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-snc6x" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.489953 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-67ptx" event={"ID":"b8150552-4914-4e92-b041-56600ed6d2f5","Type":"ContainerDied","Data":"b740ba77241ab7ac59567dc96b03935ff231d4a2d8855de4a6ed59ddebaef76b"} Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.490202 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-67ptx" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.508923 4948 scope.go:117] "RemoveContainer" containerID="e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.529465 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-snc6x"] Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.533309 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-snc6x"] Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.536380 4948 scope.go:117] "RemoveContainer" containerID="5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.540906 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-67ptx"] Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.553682 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-67ptx"] Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.564881 4948 scope.go:117] "RemoveContainer" containerID="59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e" Feb 20 08:09:47 crc kubenswrapper[4948]: E0220 08:09:47.565461 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e\": container with ID starting with 59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e not found: ID does not exist" containerID="59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.565502 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e"} err="failed to get container status \"59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e\": rpc error: code = NotFound desc = could not find container \"59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e\": container with ID starting with 59bc680a37628dd6760eae32b4fd77ee425353925297c0763e3ca5bf6107c49e not found: ID does not exist" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.565529 4948 scope.go:117] "RemoveContainer" containerID="e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e" Feb 20 08:09:47 crc kubenswrapper[4948]: E0220 08:09:47.565866 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e\": container with ID starting with e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e not found: ID does not exist" containerID="e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.565894 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e"} err="failed to get container status \"e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e\": rpc error: code = NotFound desc = could not find container \"e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e\": container with ID starting with e970f5a271f9d8c2873e14eacbf1c841136223bb57e91d7713eb158290760f1e not found: ID does not exist" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.565911 4948 scope.go:117] "RemoveContainer" containerID="5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320" Feb 20 08:09:47 crc kubenswrapper[4948]: E0220 08:09:47.566257 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320\": container with ID starting with 5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320 not found: ID does not exist" containerID="5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.566306 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320"} err="failed to get container status \"5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320\": rpc error: code = NotFound desc = could not find container \"5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320\": container with ID starting with 5e7272299ae3bd4fd920b69a3700514cf14947047239ae7f3143eafd6f2d9320 not found: ID does not exist" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.566335 4948 scope.go:117] "RemoveContainer" containerID="3ece74a70c863ce99b2f2f650f2d87de55dd27277341567aa5ed99e67f8a4dc8" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.582382 4948 scope.go:117] "RemoveContainer" containerID="e9c973e2d8ac9c5ce88b599645da506dbb8f6dbaabbec1d3e5fca44ff09a3d5c" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.606655 4948 scope.go:117] "RemoveContainer" containerID="24232a1b4e66920a8c674176ee38da24ffd8d17de7ae8a51e030daddaaeb552d" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.734328 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" path="/var/lib/kubelet/pods/b5ec9802-aa9c-4518-b99a-1673ab75ec50/volumes" Feb 20 08:09:47 crc kubenswrapper[4948]: I0220 08:09:47.735670 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" path="/var/lib/kubelet/pods/b8150552-4914-4e92-b041-56600ed6d2f5/volumes" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.967758 4948 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968592 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="extract-utilities" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968607 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="extract-utilities" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968618 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerName="extract-content" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968625 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerName="extract-content" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968638 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" containerName="extract-content" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968648 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" containerName="extract-content" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968661 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerName="extract-utilities" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968670 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerName="extract-utilities" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968685 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="extract-content" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968693 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="extract-content" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968704 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" containerName="extract-utilities" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968711 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" containerName="extract-utilities" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968721 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968728 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968739 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968747 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.968758 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968765 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968886 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d696cd-f5cf-47e0-af8d-4d20e93bd80d" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968904 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8150552-4914-4e92-b041-56600ed6d2f5" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.968913 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5ec9802-aa9c-4518-b99a-1673ab75ec50" containerName="registry-server" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.969289 4948 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.969560 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85" gracePeriod=15 Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.969723 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.970082 4948 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.970107 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244" gracePeriod=15 Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.970205 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61" gracePeriod=15 Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.970173 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556" gracePeriod=15 Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978080 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a" gracePeriod=15 Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.978190 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978221 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.978241 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978253 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.978268 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978277 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.978292 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978300 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.978310 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978319 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.978330 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978343 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 20 08:09:54 crc kubenswrapper[4948]: E0220 08:09:54.978357 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978375 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978569 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978580 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978597 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978613 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978625 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978813 4948 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Feb 20 08:09:54 crc kubenswrapper[4948]: I0220 08:09:54.978847 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.007587 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.018698 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.018748 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.018776 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.018818 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.018863 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.018896 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.018938 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.018962 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.126909 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127032 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127064 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127097 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127096 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127283 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127319 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127326 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127366 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127368 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127285 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127454 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127576 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127759 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127796 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.127817 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.303549 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:09:55 crc kubenswrapper[4948]: W0220 08:09:55.328259 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-c38fe37596ef5bf3d284b1698f5dcb7bfb184524c0da47f1359c8cf118060e1e WatchSource:0}: Error finding container c38fe37596ef5bf3d284b1698f5dcb7bfb184524c0da47f1359c8cf118060e1e: Status 404 returned error can't find the container with id c38fe37596ef5bf3d284b1698f5dcb7bfb184524c0da47f1359c8cf118060e1e Feb 20 08:09:55 crc kubenswrapper[4948]: E0220 08:09:55.331236 4948 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.113:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1895e61271a3af54 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-20 08:09:55.330510676 +0000 UTC m=+244.305005496,LastTimestamp:2026-02-20 08:09:55.330510676 +0000 UTC m=+244.305005496,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.537450 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"c38fe37596ef5bf3d284b1698f5dcb7bfb184524c0da47f1359c8cf118060e1e"} Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.540507 4948 generic.go:334] "Generic (PLEG): container finished" podID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" containerID="308dd2b5632b47b59d9d195703329e9e330103427c0967b20ee97ba81766b1b3" exitCode=0 Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.540734 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea","Type":"ContainerDied","Data":"308dd2b5632b47b59d9d195703329e9e330103427c0967b20ee97ba81766b1b3"} Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.541601 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.542209 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.544335 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.546177 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.547261 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61" exitCode=0 Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.547288 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244" exitCode=0 Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.547300 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556" exitCode=0 Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.547309 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a" exitCode=2 Feb 20 08:09:55 crc kubenswrapper[4948]: I0220 08:09:55.547350 4948 scope.go:117] "RemoveContainer" containerID="65c82ab3ae25a387b1d5c5420b22d35bce4f562e1ee57abf1c773a1850a5f678" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.557671 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47"} Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.559110 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.559526 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.561543 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.815020 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.815849 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.816232 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.955300 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-var-lock\") pod \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.955376 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kubelet-dir\") pod \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.955435 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kube-api-access\") pod \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\" (UID: \"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea\") " Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.955440 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-var-lock" (OuterVolumeSpecName: "var-lock") pod "a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" (UID: "a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.955603 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" (UID: "a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.955816 4948 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-var-lock\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.955836 4948 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:56 crc kubenswrapper[4948]: I0220 08:09:56.965087 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" (UID: "a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.056551 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.367923 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.369119 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.370296 4948 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.370814 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.371316 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.562764 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.562840 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.562876 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.562913 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.562931 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.562999 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.564628 4948 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.564682 4948 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.564729 4948 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.570057 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea","Type":"ContainerDied","Data":"501f50c7bcee3c3ff4d799d7aadad96a2760427edf1eb9577fa78b654115cf07"} Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.570115 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="501f50c7bcee3c3ff4d799d7aadad96a2760427edf1eb9577fa78b654115cf07" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.570075 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.575816 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.577205 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85" exitCode=0 Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.577304 4948 scope.go:117] "RemoveContainer" containerID="ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.577360 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.594581 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.595019 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.595669 4948 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.599196 4948 scope.go:117] "RemoveContainer" containerID="fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.604063 4948 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.604490 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.604727 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.616641 4948 scope.go:117] "RemoveContainer" containerID="b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.633894 4948 scope.go:117] "RemoveContainer" containerID="3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.647761 4948 scope.go:117] "RemoveContainer" containerID="16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.665329 4948 scope.go:117] "RemoveContainer" containerID="1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.690792 4948 scope.go:117] "RemoveContainer" containerID="ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61" Feb 20 08:09:57 crc kubenswrapper[4948]: E0220 08:09:57.697866 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\": container with ID starting with ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61 not found: ID does not exist" containerID="ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.697912 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61"} err="failed to get container status \"ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\": rpc error: code = NotFound desc = could not find container \"ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61\": container with ID starting with ff8789498468efa1fa63874e29176f0a733e428db54a0edec6902e651dc19f61 not found: ID does not exist" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.697950 4948 scope.go:117] "RemoveContainer" containerID="fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244" Feb 20 08:09:57 crc kubenswrapper[4948]: E0220 08:09:57.699081 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\": container with ID starting with fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244 not found: ID does not exist" containerID="fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.699111 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244"} err="failed to get container status \"fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\": rpc error: code = NotFound desc = could not find container \"fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244\": container with ID starting with fc9df42ce8a334551ea3179d8eabde141ffda8fd985084796b98f2cd8ee12244 not found: ID does not exist" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.699131 4948 scope.go:117] "RemoveContainer" containerID="b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556" Feb 20 08:09:57 crc kubenswrapper[4948]: E0220 08:09:57.699490 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\": container with ID starting with b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556 not found: ID does not exist" containerID="b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.699514 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556"} err="failed to get container status \"b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\": rpc error: code = NotFound desc = could not find container \"b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556\": container with ID starting with b31eba206bb7611b1a67ff9f64777879775d0256c41baa9bccf5d73d2993d556 not found: ID does not exist" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.699532 4948 scope.go:117] "RemoveContainer" containerID="3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a" Feb 20 08:09:57 crc kubenswrapper[4948]: E0220 08:09:57.699890 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\": container with ID starting with 3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a not found: ID does not exist" containerID="3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.699914 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a"} err="failed to get container status \"3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\": rpc error: code = NotFound desc = could not find container \"3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a\": container with ID starting with 3d5519390a880a1b4c9e562d6da6e91ec0aa9b33fc1bc642aceb72fbc6dd4d7a not found: ID does not exist" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.699934 4948 scope.go:117] "RemoveContainer" containerID="16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85" Feb 20 08:09:57 crc kubenswrapper[4948]: E0220 08:09:57.700389 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\": container with ID starting with 16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85 not found: ID does not exist" containerID="16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.700415 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85"} err="failed to get container status \"16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\": rpc error: code = NotFound desc = could not find container \"16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85\": container with ID starting with 16cd08df37c071e5a92d3e2c5f494a2bf4ad202fb4b4786b9cd5438a59b4de85 not found: ID does not exist" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.700432 4948 scope.go:117] "RemoveContainer" containerID="1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b" Feb 20 08:09:57 crc kubenswrapper[4948]: E0220 08:09:57.700875 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\": container with ID starting with 1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b not found: ID does not exist" containerID="1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.700901 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b"} err="failed to get container status \"1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\": rpc error: code = NotFound desc = could not find container \"1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b\": container with ID starting with 1194d72d603466c9b4434ad88cba6d7ea19d7643cd2264c9645d834c09201d5b not found: ID does not exist" Feb 20 08:09:57 crc kubenswrapper[4948]: I0220 08:09:57.732045 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 20 08:09:59 crc kubenswrapper[4948]: E0220 08:09:59.535605 4948 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.113:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1895e61271a3af54 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-20 08:09:55.330510676 +0000 UTC m=+244.305005496,LastTimestamp:2026-02-20 08:09:55.330510676 +0000 UTC m=+244.305005496,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 20 08:10:01 crc kubenswrapper[4948]: I0220 08:10:01.725272 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:01 crc kubenswrapper[4948]: I0220 08:10:01.725522 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:03 crc kubenswrapper[4948]: E0220 08:10:03.246370 4948 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:03 crc kubenswrapper[4948]: E0220 08:10:03.247603 4948 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:03 crc kubenswrapper[4948]: E0220 08:10:03.248481 4948 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:03 crc kubenswrapper[4948]: E0220 08:10:03.248964 4948 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:03 crc kubenswrapper[4948]: E0220 08:10:03.249516 4948 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:03 crc kubenswrapper[4948]: I0220 08:10:03.249572 4948 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 20 08:10:03 crc kubenswrapper[4948]: E0220 08:10:03.250044 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="200ms" Feb 20 08:10:03 crc kubenswrapper[4948]: E0220 08:10:03.450877 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="400ms" Feb 20 08:10:03 crc kubenswrapper[4948]: E0220 08:10:03.852566 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="800ms" Feb 20 08:10:04 crc kubenswrapper[4948]: E0220 08:10:04.653402 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="1.6s" Feb 20 08:10:06 crc kubenswrapper[4948]: E0220 08:10:06.254096 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.113:6443: connect: connection refused" interval="3.2s" Feb 20 08:10:06 crc kubenswrapper[4948]: I0220 08:10:06.722295 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:06 crc kubenswrapper[4948]: I0220 08:10:06.723388 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:06 crc kubenswrapper[4948]: I0220 08:10:06.723700 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:06 crc kubenswrapper[4948]: I0220 08:10:06.751171 4948 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:06 crc kubenswrapper[4948]: I0220 08:10:06.751219 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:06 crc kubenswrapper[4948]: E0220 08:10:06.751577 4948 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:06 crc kubenswrapper[4948]: I0220 08:10:06.752197 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:07 crc kubenswrapper[4948]: I0220 08:10:07.638745 4948 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="cf3409460deb96ddc07d331994e6e633d6585d5115e12c4a1dc7cdd400df8864" exitCode=0 Feb 20 08:10:07 crc kubenswrapper[4948]: I0220 08:10:07.639061 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"cf3409460deb96ddc07d331994e6e633d6585d5115e12c4a1dc7cdd400df8864"} Feb 20 08:10:07 crc kubenswrapper[4948]: I0220 08:10:07.639099 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b459704b071ca4989d633d704fc1c963dfa6f5815ee2bbd817deffc4f13de188"} Feb 20 08:10:07 crc kubenswrapper[4948]: I0220 08:10:07.639427 4948 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:07 crc kubenswrapper[4948]: I0220 08:10:07.639447 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:07 crc kubenswrapper[4948]: E0220 08:10:07.640320 4948 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:07 crc kubenswrapper[4948]: I0220 08:10:07.640424 4948 status_manager.go:851] "Failed to get status for pod" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:07 crc kubenswrapper[4948]: I0220 08:10:07.642255 4948 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.113:6443: connect: connection refused" Feb 20 08:10:08 crc kubenswrapper[4948]: I0220 08:10:08.658504 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d146155213dd15f0975d2776ef634cc79eded90bb9335fff45469a3710315ccf"} Feb 20 08:10:08 crc kubenswrapper[4948]: I0220 08:10:08.659012 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"cfdeaaf2a6e0394e1775d2d9858b907df2a41e8764c775426f1360d3a27b4a88"} Feb 20 08:10:08 crc kubenswrapper[4948]: I0220 08:10:08.659023 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fd7dd20c32f883e583e8daaaa5e8a02aaf07174aa43d4c207aa9a8b98bfee476"} Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.666753 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8b38b70f3c4ebf804f8bcc0ad1c39d452ca909f1a7cf2c456546f3d0743eb2ce"} Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.667095 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.667098 4948 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.667128 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.667107 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fb540300e2f72729de60323490ef76783142a419cc9984ab599377a9b2414525"} Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.670616 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.670740 4948 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4" exitCode=1 Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.670815 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4"} Feb 20 08:10:09 crc kubenswrapper[4948]: I0220 08:10:09.671288 4948 scope.go:117] "RemoveContainer" containerID="493d53466384aaddb8971001f61fc5e8cf0977a58dbbf48504f09c11b4db72d4" Feb 20 08:10:10 crc kubenswrapper[4948]: I0220 08:10:10.677924 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 20 08:10:10 crc kubenswrapper[4948]: I0220 08:10:10.678005 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d2fe425396bbcb915a7989ce92d49c39314fbf78eae93aca7ab82235535ad6f5"} Feb 20 08:10:11 crc kubenswrapper[4948]: I0220 08:10:11.254879 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:10:11 crc kubenswrapper[4948]: I0220 08:10:11.753553 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:11 crc kubenswrapper[4948]: I0220 08:10:11.753907 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:11 crc kubenswrapper[4948]: I0220 08:10:11.762904 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:14 crc kubenswrapper[4948]: I0220 08:10:14.678729 4948 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:14 crc kubenswrapper[4948]: I0220 08:10:14.706110 4948 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:14 crc kubenswrapper[4948]: I0220 08:10:14.706148 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:14 crc kubenswrapper[4948]: I0220 08:10:14.714398 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:14 crc kubenswrapper[4948]: I0220 08:10:14.773845 4948 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="dd37c39b-433d-4d22-aa05-3bf01b5813fe" Feb 20 08:10:14 crc kubenswrapper[4948]: I0220 08:10:14.970303 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:10:14 crc kubenswrapper[4948]: I0220 08:10:14.973783 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:10:15 crc kubenswrapper[4948]: I0220 08:10:15.713380 4948 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:15 crc kubenswrapper[4948]: I0220 08:10:15.713420 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4bc8c9ea-350e-45e5-af21-f268fb6cbffd" Feb 20 08:10:15 crc kubenswrapper[4948]: I0220 08:10:15.718861 4948 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="dd37c39b-433d-4d22-aa05-3bf01b5813fe" Feb 20 08:10:21 crc kubenswrapper[4948]: I0220 08:10:21.262879 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 08:10:24 crc kubenswrapper[4948]: I0220 08:10:24.437752 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 20 08:10:25 crc kubenswrapper[4948]: I0220 08:10:25.424337 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 20 08:10:25 crc kubenswrapper[4948]: I0220 08:10:25.796461 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 20 08:10:25 crc kubenswrapper[4948]: I0220 08:10:25.980069 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 20 08:10:26 crc kubenswrapper[4948]: I0220 08:10:26.215472 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 20 08:10:26 crc kubenswrapper[4948]: I0220 08:10:26.262062 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 20 08:10:26 crc kubenswrapper[4948]: I0220 08:10:26.344362 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 20 08:10:26 crc kubenswrapper[4948]: I0220 08:10:26.499819 4948 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 20 08:10:26 crc kubenswrapper[4948]: I0220 08:10:26.802384 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 20 08:10:26 crc kubenswrapper[4948]: I0220 08:10:26.846803 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 20 08:10:27 crc kubenswrapper[4948]: I0220 08:10:27.024161 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 20 08:10:27 crc kubenswrapper[4948]: I0220 08:10:27.024503 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 20 08:10:27 crc kubenswrapper[4948]: I0220 08:10:27.283966 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 20 08:10:27 crc kubenswrapper[4948]: I0220 08:10:27.710563 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 20 08:10:27 crc kubenswrapper[4948]: I0220 08:10:27.745821 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 20 08:10:27 crc kubenswrapper[4948]: I0220 08:10:27.769348 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 20 08:10:27 crc kubenswrapper[4948]: I0220 08:10:27.969627 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.152916 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.160730 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.225312 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.255887 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.364838 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.490718 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.615037 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.627027 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.701322 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.829250 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.885207 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 20 08:10:28 crc kubenswrapper[4948]: I0220 08:10:28.988261 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.020194 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.030615 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.140142 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.187003 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.190649 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.269838 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.384116 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.388528 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.418782 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.523323 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.632662 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.757396 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.812752 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.958845 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 20 08:10:29 crc kubenswrapper[4948]: I0220 08:10:29.982532 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.037122 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.129737 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.159252 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.286953 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.323060 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.339277 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.364709 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.453608 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.471895 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.683431 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 20 08:10:30 crc kubenswrapper[4948]: I0220 08:10:30.707998 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.088697 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.099003 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.142022 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.175650 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.240430 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.324698 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.359883 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.458098 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.533131 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.541370 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.560504 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.587631 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.622136 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.630156 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.726950 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.732406 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.800728 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.929772 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.994756 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 20 08:10:31 crc kubenswrapper[4948]: I0220 08:10:31.999901 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.116050 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.187827 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.197600 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.199565 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.216050 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.290106 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.294618 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.341361 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.352656 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.429073 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.515230 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.636898 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.694724 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.717833 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.775956 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.918805 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.962093 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.963837 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 20 08:10:32 crc kubenswrapper[4948]: I0220 08:10:32.973752 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.167897 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.170726 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.240062 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.299781 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.310708 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.363302 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.368475 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.403932 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.545097 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.562661 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.679161 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.716370 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.852172 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.891157 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 20 08:10:33 crc kubenswrapper[4948]: I0220 08:10:33.900049 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.019988 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.060469 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.122495 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.200219 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.208492 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.221587 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.307244 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.310843 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.330932 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.357798 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.364653 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.409812 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.455752 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.474355 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.566512 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.673849 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.705399 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.706100 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.849066 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.853987 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.854934 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.894550 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.897774 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.911057 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 20 08:10:34 crc kubenswrapper[4948]: I0220 08:10:34.975683 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.029162 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.127953 4948 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.131628 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.188138 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.205617 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.213716 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.243451 4948 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.249248 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=41.24921699 podStartE2EDuration="41.24921699s" podCreationTimestamp="2026-02-20 08:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:10:14.729957127 +0000 UTC m=+263.704451987" watchObservedRunningTime="2026-02-20 08:10:35.24921699 +0000 UTC m=+284.223711810" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.251614 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.253302 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.253375 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.277036 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.277005192 podStartE2EDuration="21.277005192s" podCreationTimestamp="2026-02-20 08:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:10:35.276289474 +0000 UTC m=+284.250784304" watchObservedRunningTime="2026-02-20 08:10:35.277005192 +0000 UTC m=+284.251500042" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.306696 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.350956 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.376296 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.392327 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.400903 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.459217 4948 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.462266 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.525945 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.526515 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.591485 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.592350 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.594073 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.677857 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.748035 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.773585 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 20 08:10:35 crc kubenswrapper[4948]: I0220 08:10:35.840124 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.076162 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.117214 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.156611 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.277186 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.277298 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.329247 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.338800 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.369185 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.396323 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.408452 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.491020 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.555592 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.564956 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.659437 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.696127 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.717185 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.759805 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.775891 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.812825 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.882830 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 20 08:10:36 crc kubenswrapper[4948]: I0220 08:10:36.883804 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.002211 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.013915 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.040646 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.055578 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.097206 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.107390 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.165690 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.170251 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.172215 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.306023 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.342620 4948 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.342923 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47" gracePeriod=5 Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.348251 4948 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.456201 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.481828 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.693955 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.788890 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 20 08:10:37 crc kubenswrapper[4948]: I0220 08:10:37.849490 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.100639 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.154676 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.191527 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.202256 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.203174 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.286744 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.379135 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.453529 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.469400 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.549234 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.617320 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.620229 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.672865 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.845388 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.871067 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.910714 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 20 08:10:38 crc kubenswrapper[4948]: I0220 08:10:38.950128 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.026824 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nrhl6"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.027084 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nrhl6" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerName="registry-server" containerID="cri-o://51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b" gracePeriod=30 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.046320 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n8d7m"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.046655 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n8d7m" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerName="registry-server" containerID="cri-o://acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502" gracePeriod=30 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.054164 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sjv8t"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.054679 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" podUID="95a270e5-7a08-4238-a65d-96986f2d2c39" containerName="marketplace-operator" containerID="cri-o://15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65" gracePeriod=30 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.066041 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9z42"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.066269 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c9z42" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerName="registry-server" containerID="cri-o://1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0" gracePeriod=30 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.072277 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-frznz"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.073426 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-frznz" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="registry-server" containerID="cri-o://fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd" gracePeriod=30 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.110275 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tbqzs"] Feb 20 08:10:39 crc kubenswrapper[4948]: E0220 08:10:39.110573 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" containerName="installer" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.110602 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" containerName="installer" Feb 20 08:10:39 crc kubenswrapper[4948]: E0220 08:10:39.110663 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.110677 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.110816 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.110849 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0fc48a9-8f2e-4a58-9ec2-0ab67f6cedea" containerName="installer" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.111467 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.116316 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.123175 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tbqzs"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.185083 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.222385 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.225387 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1e3d2cd4-4311-4454-b257-20a3caf243b3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.225440 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1e3d2cd4-4311-4454-b257-20a3caf243b3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.225460 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfqrz\" (UniqueName: \"kubernetes.io/projected/1e3d2cd4-4311-4454-b257-20a3caf243b3-kube-api-access-cfqrz\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.327083 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1e3d2cd4-4311-4454-b257-20a3caf243b3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.327129 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1e3d2cd4-4311-4454-b257-20a3caf243b3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.327154 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfqrz\" (UniqueName: \"kubernetes.io/projected/1e3d2cd4-4311-4454-b257-20a3caf243b3-kube-api-access-cfqrz\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.328432 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1e3d2cd4-4311-4454-b257-20a3caf243b3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.334798 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1e3d2cd4-4311-4454-b257-20a3caf243b3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.340629 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.343516 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfqrz\" (UniqueName: \"kubernetes.io/projected/1e3d2cd4-4311-4454-b257-20a3caf243b3-kube-api-access-cfqrz\") pod \"marketplace-operator-79b997595-tbqzs\" (UID: \"1e3d2cd4-4311-4454-b257-20a3caf243b3\") " pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.354171 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.373929 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.452836 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.464328 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.502785 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.508812 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.512671 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.517569 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.581243 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.593418 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.616421 4948 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.626022 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632123 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-utilities\") pod \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632178 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55q6w\" (UniqueName: \"kubernetes.io/projected/95a270e5-7a08-4238-a65d-96986f2d2c39-kube-api-access-55q6w\") pod \"95a270e5-7a08-4238-a65d-96986f2d2c39\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632211 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxq9k\" (UniqueName: \"kubernetes.io/projected/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-kube-api-access-kxq9k\") pod \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632231 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-catalog-content\") pod \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632250 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqtvv\" (UniqueName: \"kubernetes.io/projected/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-kube-api-access-cqtvv\") pod \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632267 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-catalog-content\") pod \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\" (UID: \"c53ddd33-0f9f-4794-b346-7d48a6c09c9b\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632286 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-catalog-content\") pod \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632322 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-utilities\") pod \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632342 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-utilities\") pod \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632356 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqs4x\" (UniqueName: \"kubernetes.io/projected/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-kube-api-access-hqs4x\") pod \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\" (UID: \"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632374 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-operator-metrics\") pod \"95a270e5-7a08-4238-a65d-96986f2d2c39\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632413 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc669\" (UniqueName: \"kubernetes.io/projected/f75a50cf-7dae-420b-a00d-671ede3cb6f7-kube-api-access-nc669\") pod \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632443 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-catalog-content\") pod \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\" (UID: \"f75a50cf-7dae-420b-a00d-671ede3cb6f7\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632465 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-trusted-ca\") pod \"95a270e5-7a08-4238-a65d-96986f2d2c39\" (UID: \"95a270e5-7a08-4238-a65d-96986f2d2c39\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.632481 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-utilities\") pod \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\" (UID: \"00f4b4c1-af36-4ccd-a3d1-dd6df1186338\") " Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.633587 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-utilities" (OuterVolumeSpecName: "utilities") pod "c53ddd33-0f9f-4794-b346-7d48a6c09c9b" (UID: "c53ddd33-0f9f-4794-b346-7d48a6c09c9b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.634332 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-utilities" (OuterVolumeSpecName: "utilities") pod "dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" (UID: "dc2fe1f3-084a-4e45-b64d-f385fb28d7e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.634573 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-utilities" (OuterVolumeSpecName: "utilities") pod "f75a50cf-7dae-420b-a00d-671ede3cb6f7" (UID: "f75a50cf-7dae-420b-a00d-671ede3cb6f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.635357 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "95a270e5-7a08-4238-a65d-96986f2d2c39" (UID: "95a270e5-7a08-4238-a65d-96986f2d2c39"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.635452 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-utilities" (OuterVolumeSpecName: "utilities") pod "00f4b4c1-af36-4ccd-a3d1-dd6df1186338" (UID: "00f4b4c1-af36-4ccd-a3d1-dd6df1186338"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.638495 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f75a50cf-7dae-420b-a00d-671ede3cb6f7-kube-api-access-nc669" (OuterVolumeSpecName: "kube-api-access-nc669") pod "f75a50cf-7dae-420b-a00d-671ede3cb6f7" (UID: "f75a50cf-7dae-420b-a00d-671ede3cb6f7"). InnerVolumeSpecName "kube-api-access-nc669". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.638688 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-kube-api-access-cqtvv" (OuterVolumeSpecName: "kube-api-access-cqtvv") pod "c53ddd33-0f9f-4794-b346-7d48a6c09c9b" (UID: "c53ddd33-0f9f-4794-b346-7d48a6c09c9b"). InnerVolumeSpecName "kube-api-access-cqtvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.638821 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-kube-api-access-hqs4x" (OuterVolumeSpecName: "kube-api-access-hqs4x") pod "dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" (UID: "dc2fe1f3-084a-4e45-b64d-f385fb28d7e9"). InnerVolumeSpecName "kube-api-access-hqs4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.641780 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-kube-api-access-kxq9k" (OuterVolumeSpecName: "kube-api-access-kxq9k") pod "00f4b4c1-af36-4ccd-a3d1-dd6df1186338" (UID: "00f4b4c1-af36-4ccd-a3d1-dd6df1186338"). InnerVolumeSpecName "kube-api-access-kxq9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.651549 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "95a270e5-7a08-4238-a65d-96986f2d2c39" (UID: "95a270e5-7a08-4238-a65d-96986f2d2c39"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.662348 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95a270e5-7a08-4238-a65d-96986f2d2c39-kube-api-access-55q6w" (OuterVolumeSpecName: "kube-api-access-55q6w") pod "95a270e5-7a08-4238-a65d-96986f2d2c39" (UID: "95a270e5-7a08-4238-a65d-96986f2d2c39"). InnerVolumeSpecName "kube-api-access-55q6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.664879 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c53ddd33-0f9f-4794-b346-7d48a6c09c9b" (UID: "c53ddd33-0f9f-4794-b346-7d48a6c09c9b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.676199 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.689313 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tbqzs"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.724894 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00f4b4c1-af36-4ccd-a3d1-dd6df1186338" (UID: "00f4b4c1-af36-4ccd-a3d1-dd6df1186338"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733434 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc669\" (UniqueName: \"kubernetes.io/projected/f75a50cf-7dae-420b-a00d-671ede3cb6f7-kube-api-access-nc669\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733767 4948 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733777 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733787 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733799 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55q6w\" (UniqueName: \"kubernetes.io/projected/95a270e5-7a08-4238-a65d-96986f2d2c39-kube-api-access-55q6w\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733809 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxq9k\" (UniqueName: \"kubernetes.io/projected/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-kube-api-access-kxq9k\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733818 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f4b4c1-af36-4ccd-a3d1-dd6df1186338-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733828 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733836 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqtvv\" (UniqueName: \"kubernetes.io/projected/c53ddd33-0f9f-4794-b346-7d48a6c09c9b-kube-api-access-cqtvv\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733844 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733853 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733861 4948 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95a270e5-7a08-4238-a65d-96986f2d2c39-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.733891 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqs4x\" (UniqueName: \"kubernetes.io/projected/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-kube-api-access-hqs4x\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.738284 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f75a50cf-7dae-420b-a00d-671ede3cb6f7" (UID: "f75a50cf-7dae-420b-a00d-671ede3cb6f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.780533 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" (UID: "dc2fe1f3-084a-4e45-b64d-f385fb28d7e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.835502 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.835601 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f75a50cf-7dae-420b-a00d-671ede3cb6f7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.862792 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.875741 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.885233 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.885326 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9z42" event={"ID":"c53ddd33-0f9f-4794-b346-7d48a6c09c9b","Type":"ContainerDied","Data":"1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.885369 4948 scope.go:117] "RemoveContainer" containerID="1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.885372 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9z42" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.885299 4948 generic.go:334] "Generic (PLEG): container finished" podID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerID="1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0" exitCode=0 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.885477 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9z42" event={"ID":"c53ddd33-0f9f-4794-b346-7d48a6c09c9b","Type":"ContainerDied","Data":"9a7c5f1538f439aba0332cfbd20f5f653d7b240472d961eff61974598f86365b"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.890474 4948 generic.go:334] "Generic (PLEG): container finished" podID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerID="fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd" exitCode=0 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.890539 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frznz" event={"ID":"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9","Type":"ContainerDied","Data":"fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.890568 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frznz" event={"ID":"dc2fe1f3-084a-4e45-b64d-f385fb28d7e9","Type":"ContainerDied","Data":"bc370df4b0b552966157b6ac4af296601fb860ed2508da9c4291e9924ec1e9cb"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.890626 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frznz" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.897407 4948 generic.go:334] "Generic (PLEG): container finished" podID="95a270e5-7a08-4238-a65d-96986f2d2c39" containerID="15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65" exitCode=0 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.897671 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.899224 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" event={"ID":"95a270e5-7a08-4238-a65d-96986f2d2c39","Type":"ContainerDied","Data":"15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.899306 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sjv8t" event={"ID":"95a270e5-7a08-4238-a65d-96986f2d2c39","Type":"ContainerDied","Data":"936301da215b82b117fb711a926cf169f025d16470e0871be5907e8efa5aec8e"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.903654 4948 generic.go:334] "Generic (PLEG): container finished" podID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerID="acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502" exitCode=0 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.903738 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8d7m" event={"ID":"00f4b4c1-af36-4ccd-a3d1-dd6df1186338","Type":"ContainerDied","Data":"acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.903769 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8d7m" event={"ID":"00f4b4c1-af36-4ccd-a3d1-dd6df1186338","Type":"ContainerDied","Data":"33b053c8152742a922c7d8709831de18b8993625ae7594197966e98f173b1174"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.903848 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n8d7m" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.910309 4948 generic.go:334] "Generic (PLEG): container finished" podID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerID="51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b" exitCode=0 Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.910374 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrhl6" event={"ID":"f75a50cf-7dae-420b-a00d-671ede3cb6f7","Type":"ContainerDied","Data":"51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.910400 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrhl6" event={"ID":"f75a50cf-7dae-420b-a00d-671ede3cb6f7","Type":"ContainerDied","Data":"d0015edf4c699a29515d340ae2c01786195a0e9d51c4459eae81be82178c452a"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.910459 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrhl6" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.912810 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" event={"ID":"1e3d2cd4-4311-4454-b257-20a3caf243b3","Type":"ContainerStarted","Data":"313c18eb1c126858c938cbb0519630cfedb0f11cc4071cb91007fa8807d62753"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.912849 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" event={"ID":"1e3d2cd4-4311-4454-b257-20a3caf243b3","Type":"ContainerStarted","Data":"60097ca12a5b208298fe3999bd1e665bf320413d781422c5a1f9dc265d840874"} Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.913773 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.920137 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9z42"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.920187 4948 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tbqzs container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" start-of-body= Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.920267 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" podUID="1e3d2cd4-4311-4454-b257-20a3caf243b3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.920850 4948 scope.go:117] "RemoveContainer" containerID="bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.935414 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9z42"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.939130 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sjv8t"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.941921 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sjv8t"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.953532 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" podStartSLOduration=0.953514816 podStartE2EDuration="953.514816ms" podCreationTimestamp="2026-02-20 08:10:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:10:39.951504847 +0000 UTC m=+288.925999677" watchObservedRunningTime="2026-02-20 08:10:39.953514816 +0000 UTC m=+288.928009646" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.963187 4948 scope.go:117] "RemoveContainer" containerID="1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.971762 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n8d7m"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.977914 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n8d7m"] Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.986734 4948 scope.go:117] "RemoveContainer" containerID="1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.987655 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-frznz"] Feb 20 08:10:39 crc kubenswrapper[4948]: E0220 08:10:39.988381 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0\": container with ID starting with 1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0 not found: ID does not exist" containerID="1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.988443 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0"} err="failed to get container status \"1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0\": rpc error: code = NotFound desc = could not find container \"1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0\": container with ID starting with 1c4c2fe97c46c10efdb2a7e794505136dc520212167c78787ccbac94170a83c0 not found: ID does not exist" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.988475 4948 scope.go:117] "RemoveContainer" containerID="bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83" Feb 20 08:10:39 crc kubenswrapper[4948]: E0220 08:10:39.989320 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83\": container with ID starting with bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83 not found: ID does not exist" containerID="bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.989407 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83"} err="failed to get container status \"bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83\": rpc error: code = NotFound desc = could not find container \"bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83\": container with ID starting with bfea6ab4eabff6af6e0a3b30d304544693e06a2189af921dbd30d2b97a616f83 not found: ID does not exist" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.989492 4948 scope.go:117] "RemoveContainer" containerID="1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe" Feb 20 08:10:39 crc kubenswrapper[4948]: E0220 08:10:39.990158 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe\": container with ID starting with 1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe not found: ID does not exist" containerID="1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.990207 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe"} err="failed to get container status \"1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe\": rpc error: code = NotFound desc = could not find container \"1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe\": container with ID starting with 1d07ae65c7f4887c7cc6c699e956a80f175ba6deadc7da37863658727e044dbe not found: ID does not exist" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.990240 4948 scope.go:117] "RemoveContainer" containerID="fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd" Feb 20 08:10:39 crc kubenswrapper[4948]: I0220 08:10:39.991778 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-frznz"] Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.000250 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nrhl6"] Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.004032 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nrhl6"] Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.011459 4948 scope.go:117] "RemoveContainer" containerID="27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.025717 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.028781 4948 scope.go:117] "RemoveContainer" containerID="6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.044215 4948 scope.go:117] "RemoveContainer" containerID="fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.044819 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd\": container with ID starting with fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd not found: ID does not exist" containerID="fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.044945 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd"} err="failed to get container status \"fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd\": rpc error: code = NotFound desc = could not find container \"fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd\": container with ID starting with fb097a14ed0ab7d5af6abab587a1c01e68baf8b43ccad645f65796167b0c90bd not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.045056 4948 scope.go:117] "RemoveContainer" containerID="27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.045485 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00\": container with ID starting with 27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00 not found: ID does not exist" containerID="27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.045571 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00"} err="failed to get container status \"27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00\": rpc error: code = NotFound desc = could not find container \"27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00\": container with ID starting with 27b14a9abd1358ef732430763a33e8727bba43c949149d1b6052fc586fb0bd00 not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.045651 4948 scope.go:117] "RemoveContainer" containerID="6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.046103 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7\": container with ID starting with 6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7 not found: ID does not exist" containerID="6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.046206 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7"} err="failed to get container status \"6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7\": rpc error: code = NotFound desc = could not find container \"6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7\": container with ID starting with 6e6f9609316227fbb42f161491bc563af5bbab43a8ee401a26665bf588b0e5d7 not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.046286 4948 scope.go:117] "RemoveContainer" containerID="15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.061730 4948 scope.go:117] "RemoveContainer" containerID="15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.062723 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65\": container with ID starting with 15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65 not found: ID does not exist" containerID="15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.062762 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65"} err="failed to get container status \"15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65\": rpc error: code = NotFound desc = could not find container \"15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65\": container with ID starting with 15687151f4ed59c06cf2071d3204504981cf4a0c6cdb40d25a5dbd083313ed65 not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.062790 4948 scope.go:117] "RemoveContainer" containerID="acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.076981 4948 scope.go:117] "RemoveContainer" containerID="5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.094332 4948 scope.go:117] "RemoveContainer" containerID="8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.105866 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.110917 4948 scope.go:117] "RemoveContainer" containerID="acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.111466 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502\": container with ID starting with acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502 not found: ID does not exist" containerID="acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.111579 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502"} err="failed to get container status \"acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502\": rpc error: code = NotFound desc = could not find container \"acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502\": container with ID starting with acfb3ae1070bc662949f97e63b0d95f3a1957ddfef18898cde66fcbe8186d502 not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.111677 4948 scope.go:117] "RemoveContainer" containerID="5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.112099 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7\": container with ID starting with 5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7 not found: ID does not exist" containerID="5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.112206 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7"} err="failed to get container status \"5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7\": rpc error: code = NotFound desc = could not find container \"5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7\": container with ID starting with 5e9a0379202da2a09189213d51042628595d46ea2ac1801d0bcbd926088826d7 not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.112303 4948 scope.go:117] "RemoveContainer" containerID="8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.114046 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f\": container with ID starting with 8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f not found: ID does not exist" containerID="8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.114149 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f"} err="failed to get container status \"8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f\": rpc error: code = NotFound desc = could not find container \"8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f\": container with ID starting with 8ae4d6ef7b64ba80a0da276ad7998550243e3a26aaa0bbf3282f22cf4366103f not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.114249 4948 scope.go:117] "RemoveContainer" containerID="51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.133522 4948 scope.go:117] "RemoveContainer" containerID="ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.150603 4948 scope.go:117] "RemoveContainer" containerID="df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.152571 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.170268 4948 scope.go:117] "RemoveContainer" containerID="51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.170871 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b\": container with ID starting with 51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b not found: ID does not exist" containerID="51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.170944 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b"} err="failed to get container status \"51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b\": rpc error: code = NotFound desc = could not find container \"51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b\": container with ID starting with 51c14bb6285644fd84e90599149f7f330a4395b58ff4509817fbdce179d6470b not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.171013 4948 scope.go:117] "RemoveContainer" containerID="ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.171957 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37\": container with ID starting with ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37 not found: ID does not exist" containerID="ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.172022 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37"} err="failed to get container status \"ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37\": rpc error: code = NotFound desc = could not find container \"ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37\": container with ID starting with ed76ff8cc753dd2cd048a3406074573b902071e966e21613ec37f782fb620d37 not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.172058 4948 scope.go:117] "RemoveContainer" containerID="df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e" Feb 20 08:10:40 crc kubenswrapper[4948]: E0220 08:10:40.172485 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e\": container with ID starting with df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e not found: ID does not exist" containerID="df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.172593 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e"} err="failed to get container status \"df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e\": rpc error: code = NotFound desc = could not find container \"df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e\": container with ID starting with df3ff28a9e161bcd3a383de17aefbe231a72f4c519e331260796574dc38d393e not found: ID does not exist" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.174268 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.334873 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.349746 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.397873 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.425224 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.460868 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.584851 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.719298 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.778283 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.885936 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.904205 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.905322 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 20 08:10:40 crc kubenswrapper[4948]: I0220 08:10:40.945049 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-tbqzs" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.264455 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.331441 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.384908 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.514375 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.734810 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" path="/var/lib/kubelet/pods/00f4b4c1-af36-4ccd-a3d1-dd6df1186338/volumes" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.736222 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95a270e5-7a08-4238-a65d-96986f2d2c39" path="/var/lib/kubelet/pods/95a270e5-7a08-4238-a65d-96986f2d2c39/volumes" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.738184 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" path="/var/lib/kubelet/pods/c53ddd33-0f9f-4794-b346-7d48a6c09c9b/volumes" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.740316 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" path="/var/lib/kubelet/pods/dc2fe1f3-084a-4e45-b64d-f385fb28d7e9/volumes" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.741620 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" path="/var/lib/kubelet/pods/f75a50cf-7dae-420b-a00d-671ede3cb6f7/volumes" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.849828 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.879837 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.978863 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 20 08:10:41 crc kubenswrapper[4948]: I0220 08:10:41.990024 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 20 08:10:42 crc kubenswrapper[4948]: I0220 08:10:42.935709 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 20 08:10:42 crc kubenswrapper[4948]: I0220 08:10:42.944953 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 20 08:10:42 crc kubenswrapper[4948]: I0220 08:10:42.945110 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:10:42 crc kubenswrapper[4948]: I0220 08:10:42.980669 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 20 08:10:42 crc kubenswrapper[4948]: I0220 08:10:42.980740 4948 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47" exitCode=137 Feb 20 08:10:42 crc kubenswrapper[4948]: I0220 08:10:42.981225 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 08:10:42 crc kubenswrapper[4948]: I0220 08:10:42.981419 4948 scope.go:117] "RemoveContainer" containerID="955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.004517 4948 scope.go:117] "RemoveContainer" containerID="955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47" Feb 20 08:10:43 crc kubenswrapper[4948]: E0220 08:10:43.005792 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47\": container with ID starting with 955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47 not found: ID does not exist" containerID="955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.005851 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47"} err="failed to get container status \"955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47\": rpc error: code = NotFound desc = could not find container \"955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47\": container with ID starting with 955d8cda21a3da3121dfa54a2504d8b0f8470dbb104720ff4e180c5ad7329f47 not found: ID does not exist" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.079705 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.079744 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.079823 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.079881 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.079905 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.079931 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.079937 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.079989 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.080000 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.080337 4948 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.080358 4948 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.080372 4948 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.080381 4948 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.086895 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.158455 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.181814 4948 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.735571 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.736311 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.748915 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.748968 4948 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="872f9266-e8c5-46f8-bddb-c112fa6260a9" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.753326 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.753371 4948 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="872f9266-e8c5-46f8-bddb-c112fa6260a9" Feb 20 08:10:43 crc kubenswrapper[4948]: I0220 08:10:43.880698 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 20 08:10:51 crc kubenswrapper[4948]: I0220 08:10:51.489820 4948 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.474152 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dl7nj"] Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475029 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a270e5-7a08-4238-a65d-96986f2d2c39" containerName="marketplace-operator" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475050 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a270e5-7a08-4238-a65d-96986f2d2c39" containerName="marketplace-operator" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475073 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerName="extract-content" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475086 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerName="extract-content" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475103 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerName="extract-content" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475115 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerName="extract-content" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475130 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475141 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475161 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475172 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475190 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerName="extract-utilities" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475203 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerName="extract-utilities" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475220 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="extract-utilities" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475232 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="extract-utilities" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475251 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475263 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475276 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerName="extract-utilities" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475289 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerName="extract-utilities" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475305 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="extract-content" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475316 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="extract-content" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475334 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerName="extract-content" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475346 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerName="extract-content" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475361 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475373 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: E0220 08:11:05.475394 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerName="extract-utilities" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475405 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerName="extract-utilities" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475561 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="00f4b4c1-af36-4ccd-a3d1-dd6df1186338" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475581 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f75a50cf-7dae-420b-a00d-671ede3cb6f7" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475595 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc2fe1f3-084a-4e45-b64d-f385fb28d7e9" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475613 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="95a270e5-7a08-4238-a65d-96986f2d2c39" containerName="marketplace-operator" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.475630 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="c53ddd33-0f9f-4794-b346-7d48a6c09c9b" containerName="registry-server" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.476852 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.481192 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.489621 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e759dc6-36e8-4a97-b994-e657f1b5d540-utilities\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.489721 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqd6w\" (UniqueName: \"kubernetes.io/projected/0e759dc6-36e8-4a97-b994-e657f1b5d540-kube-api-access-lqd6w\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.489842 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e759dc6-36e8-4a97-b994-e657f1b5d540-catalog-content\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.496112 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dl7nj"] Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.591201 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e759dc6-36e8-4a97-b994-e657f1b5d540-catalog-content\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.591403 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e759dc6-36e8-4a97-b994-e657f1b5d540-utilities\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.591439 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqd6w\" (UniqueName: \"kubernetes.io/projected/0e759dc6-36e8-4a97-b994-e657f1b5d540-kube-api-access-lqd6w\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.592825 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e759dc6-36e8-4a97-b994-e657f1b5d540-utilities\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.593014 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e759dc6-36e8-4a97-b994-e657f1b5d540-catalog-content\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.618160 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqd6w\" (UniqueName: \"kubernetes.io/projected/0e759dc6-36e8-4a97-b994-e657f1b5d540-kube-api-access-lqd6w\") pod \"community-operators-dl7nj\" (UID: \"0e759dc6-36e8-4a97-b994-e657f1b5d540\") " pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.664295 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tt6t6"] Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.666334 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.669304 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.679754 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tt6t6"] Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.692515 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsp6z\" (UniqueName: \"kubernetes.io/projected/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-kube-api-access-nsp6z\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.692568 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-catalog-content\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.692633 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-utilities\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.794408 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsp6z\" (UniqueName: \"kubernetes.io/projected/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-kube-api-access-nsp6z\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.794537 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-catalog-content\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.794734 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-utilities\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.795474 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-catalog-content\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.795890 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-utilities\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.819886 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsp6z\" (UniqueName: \"kubernetes.io/projected/94e3bb0b-818a-491e-8e1e-a1664f3a9f12-kube-api-access-nsp6z\") pod \"certified-operators-tt6t6\" (UID: \"94e3bb0b-818a-491e-8e1e-a1664f3a9f12\") " pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.844666 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:05 crc kubenswrapper[4948]: I0220 08:11:05.989500 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:06 crc kubenswrapper[4948]: I0220 08:11:06.104921 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dl7nj"] Feb 20 08:11:06 crc kubenswrapper[4948]: I0220 08:11:06.122960 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dl7nj" event={"ID":"0e759dc6-36e8-4a97-b994-e657f1b5d540","Type":"ContainerStarted","Data":"bf3ed40d3cc7b0698c95b69e2698db264f0323aa1832e449fcfe404052daf599"} Feb 20 08:11:06 crc kubenswrapper[4948]: I0220 08:11:06.249219 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tt6t6"] Feb 20 08:11:06 crc kubenswrapper[4948]: W0220 08:11:06.255704 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94e3bb0b_818a_491e_8e1e_a1664f3a9f12.slice/crio-a46eccf52e321b9645691299b6786a0989ece4633f0c772e19e8616200646554 WatchSource:0}: Error finding container a46eccf52e321b9645691299b6786a0989ece4633f0c772e19e8616200646554: Status 404 returned error can't find the container with id a46eccf52e321b9645691299b6786a0989ece4633f0c772e19e8616200646554 Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.132703 4948 generic.go:334] "Generic (PLEG): container finished" podID="0e759dc6-36e8-4a97-b994-e657f1b5d540" containerID="7a7848ac580be195cccad9466f97fab3d16ed5cce26ab1658187f2cafddd55eb" exitCode=0 Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.132835 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dl7nj" event={"ID":"0e759dc6-36e8-4a97-b994-e657f1b5d540","Type":"ContainerDied","Data":"7a7848ac580be195cccad9466f97fab3d16ed5cce26ab1658187f2cafddd55eb"} Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.137267 4948 generic.go:334] "Generic (PLEG): container finished" podID="94e3bb0b-818a-491e-8e1e-a1664f3a9f12" containerID="5404601cc08a73a981959eb4e9f568980c021e72d4fb5ec5a3c470eab642d7d1" exitCode=0 Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.137354 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt6t6" event={"ID":"94e3bb0b-818a-491e-8e1e-a1664f3a9f12","Type":"ContainerDied","Data":"5404601cc08a73a981959eb4e9f568980c021e72d4fb5ec5a3c470eab642d7d1"} Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.137409 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt6t6" event={"ID":"94e3bb0b-818a-491e-8e1e-a1664f3a9f12","Type":"ContainerStarted","Data":"a46eccf52e321b9645691299b6786a0989ece4633f0c772e19e8616200646554"} Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.869709 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x7bcm"] Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.871030 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.874800 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.890816 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7bcm"] Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.934326 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ec98112-e189-4a27-a6dd-8cb035596810-utilities\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.934740 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ec98112-e189-4a27-a6dd-8cb035596810-catalog-content\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:07 crc kubenswrapper[4948]: I0220 08:11:07.934764 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bplnw\" (UniqueName: \"kubernetes.io/projected/6ec98112-e189-4a27-a6dd-8cb035596810-kube-api-access-bplnw\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.035641 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ec98112-e189-4a27-a6dd-8cb035596810-utilities\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.035756 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ec98112-e189-4a27-a6dd-8cb035596810-catalog-content\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.035792 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bplnw\" (UniqueName: \"kubernetes.io/projected/6ec98112-e189-4a27-a6dd-8cb035596810-kube-api-access-bplnw\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.036153 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ec98112-e189-4a27-a6dd-8cb035596810-utilities\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.036416 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ec98112-e189-4a27-a6dd-8cb035596810-catalog-content\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.071043 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jbtm5"] Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.073320 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.076931 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.082645 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bplnw\" (UniqueName: \"kubernetes.io/projected/6ec98112-e189-4a27-a6dd-8cb035596810-kube-api-access-bplnw\") pod \"redhat-marketplace-x7bcm\" (UID: \"6ec98112-e189-4a27-a6dd-8cb035596810\") " pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.084864 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jbtm5"] Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.136463 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zmfg\" (UniqueName: \"kubernetes.io/projected/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-kube-api-access-8zmfg\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.136517 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-catalog-content\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.136722 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-utilities\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.151942 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt6t6" event={"ID":"94e3bb0b-818a-491e-8e1e-a1664f3a9f12","Type":"ContainerStarted","Data":"f1888dbac65f5a2dda53ac9b186d1b21ab5834fdf1cf175e2d58d53cb700d38b"} Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.216364 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.238146 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zmfg\" (UniqueName: \"kubernetes.io/projected/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-kube-api-access-8zmfg\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.238204 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-catalog-content\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.238253 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-utilities\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.238872 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-utilities\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.239766 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-catalog-content\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.267711 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zmfg\" (UniqueName: \"kubernetes.io/projected/5ee242c9-9783-44cb-a4dc-1ea50a7ae185-kube-api-access-8zmfg\") pod \"redhat-operators-jbtm5\" (UID: \"5ee242c9-9783-44cb-a4dc-1ea50a7ae185\") " pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.464263 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7bcm"] Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.466657 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:08 crc kubenswrapper[4948]: W0220 08:11:08.551760 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ec98112_e189_4a27_a6dd_8cb035596810.slice/crio-3981f65f9b36579b160b94eda6a0c154957916c3d4d7e6b5d32e34517e7ed66c WatchSource:0}: Error finding container 3981f65f9b36579b160b94eda6a0c154957916c3d4d7e6b5d32e34517e7ed66c: Status 404 returned error can't find the container with id 3981f65f9b36579b160b94eda6a0c154957916c3d4d7e6b5d32e34517e7ed66c Feb 20 08:11:08 crc kubenswrapper[4948]: I0220 08:11:08.696381 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jbtm5"] Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.163297 4948 generic.go:334] "Generic (PLEG): container finished" podID="6ec98112-e189-4a27-a6dd-8cb035596810" containerID="b53992b5ed6f1bf4c2505a8d4b9914069c430e74cf847c56c06e70e9587d3fee" exitCode=0 Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.163377 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7bcm" event={"ID":"6ec98112-e189-4a27-a6dd-8cb035596810","Type":"ContainerDied","Data":"b53992b5ed6f1bf4c2505a8d4b9914069c430e74cf847c56c06e70e9587d3fee"} Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.163684 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7bcm" event={"ID":"6ec98112-e189-4a27-a6dd-8cb035596810","Type":"ContainerStarted","Data":"3981f65f9b36579b160b94eda6a0c154957916c3d4d7e6b5d32e34517e7ed66c"} Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.167111 4948 generic.go:334] "Generic (PLEG): container finished" podID="0e759dc6-36e8-4a97-b994-e657f1b5d540" containerID="1564a74185083f2b92c7307fe86058881bf28dadd08bdfcd91b901b9737c79c3" exitCode=0 Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.167241 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dl7nj" event={"ID":"0e759dc6-36e8-4a97-b994-e657f1b5d540","Type":"ContainerDied","Data":"1564a74185083f2b92c7307fe86058881bf28dadd08bdfcd91b901b9737c79c3"} Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.170783 4948 generic.go:334] "Generic (PLEG): container finished" podID="94e3bb0b-818a-491e-8e1e-a1664f3a9f12" containerID="f1888dbac65f5a2dda53ac9b186d1b21ab5834fdf1cf175e2d58d53cb700d38b" exitCode=0 Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.170876 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt6t6" event={"ID":"94e3bb0b-818a-491e-8e1e-a1664f3a9f12","Type":"ContainerDied","Data":"f1888dbac65f5a2dda53ac9b186d1b21ab5834fdf1cf175e2d58d53cb700d38b"} Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.173544 4948 generic.go:334] "Generic (PLEG): container finished" podID="5ee242c9-9783-44cb-a4dc-1ea50a7ae185" containerID="2218bda4dd124f74ba9c850581183328566b25c77cae393e864df3ccbbbabb4b" exitCode=0 Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.173598 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbtm5" event={"ID":"5ee242c9-9783-44cb-a4dc-1ea50a7ae185","Type":"ContainerDied","Data":"2218bda4dd124f74ba9c850581183328566b25c77cae393e864df3ccbbbabb4b"} Feb 20 08:11:09 crc kubenswrapper[4948]: I0220 08:11:09.173629 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbtm5" event={"ID":"5ee242c9-9783-44cb-a4dc-1ea50a7ae185","Type":"ContainerStarted","Data":"dc93ca20aabfe84b972cccb2fcd744cdb44482b26917eb4a96821537d451ddf2"} Feb 20 08:11:10 crc kubenswrapper[4948]: I0220 08:11:10.180998 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbtm5" event={"ID":"5ee242c9-9783-44cb-a4dc-1ea50a7ae185","Type":"ContainerStarted","Data":"1da87f05a71242abf8441f6ff99031f27ba7460a91f1265f6101e951bc82ac76"} Feb 20 08:11:10 crc kubenswrapper[4948]: I0220 08:11:10.182850 4948 generic.go:334] "Generic (PLEG): container finished" podID="6ec98112-e189-4a27-a6dd-8cb035596810" containerID="d7bab80a1ea22ca8894e254a51e0da2b70189f46ceebd8ec1a2d3b1dbba46468" exitCode=0 Feb 20 08:11:10 crc kubenswrapper[4948]: I0220 08:11:10.183517 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7bcm" event={"ID":"6ec98112-e189-4a27-a6dd-8cb035596810","Type":"ContainerDied","Data":"d7bab80a1ea22ca8894e254a51e0da2b70189f46ceebd8ec1a2d3b1dbba46468"} Feb 20 08:11:10 crc kubenswrapper[4948]: I0220 08:11:10.186707 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dl7nj" event={"ID":"0e759dc6-36e8-4a97-b994-e657f1b5d540","Type":"ContainerStarted","Data":"c2081b84a51e60cc8445891129a696d808670e42dc13946e716842b6c395bbad"} Feb 20 08:11:10 crc kubenswrapper[4948]: I0220 08:11:10.189006 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt6t6" event={"ID":"94e3bb0b-818a-491e-8e1e-a1664f3a9f12","Type":"ContainerStarted","Data":"e20b9b3ad805b3378f00b56f519bee3a7a9266636a96258528a39123833e60c8"} Feb 20 08:11:10 crc kubenswrapper[4948]: I0220 08:11:10.222850 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tt6t6" podStartSLOduration=2.708221386 podStartE2EDuration="5.222825693s" podCreationTimestamp="2026-02-20 08:11:05 +0000 UTC" firstStartedPulling="2026-02-20 08:11:07.141219431 +0000 UTC m=+316.115714251" lastFinishedPulling="2026-02-20 08:11:09.655823718 +0000 UTC m=+318.630318558" observedRunningTime="2026-02-20 08:11:10.218552008 +0000 UTC m=+319.193046828" watchObservedRunningTime="2026-02-20 08:11:10.222825693 +0000 UTC m=+319.197320553" Feb 20 08:11:10 crc kubenswrapper[4948]: I0220 08:11:10.238563 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dl7nj" podStartSLOduration=2.823887953 podStartE2EDuration="5.238545258s" podCreationTimestamp="2026-02-20 08:11:05 +0000 UTC" firstStartedPulling="2026-02-20 08:11:07.13504133 +0000 UTC m=+316.109536180" lastFinishedPulling="2026-02-20 08:11:09.549698665 +0000 UTC m=+318.524193485" observedRunningTime="2026-02-20 08:11:10.237156434 +0000 UTC m=+319.211651304" watchObservedRunningTime="2026-02-20 08:11:10.238545258 +0000 UTC m=+319.213040078" Feb 20 08:11:11 crc kubenswrapper[4948]: I0220 08:11:11.195933 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7bcm" event={"ID":"6ec98112-e189-4a27-a6dd-8cb035596810","Type":"ContainerStarted","Data":"fee95fffee0a6e41238f5067917808cd72cf7d7d678b2ec55baeee91c3f20830"} Feb 20 08:11:11 crc kubenswrapper[4948]: I0220 08:11:11.197936 4948 generic.go:334] "Generic (PLEG): container finished" podID="5ee242c9-9783-44cb-a4dc-1ea50a7ae185" containerID="1da87f05a71242abf8441f6ff99031f27ba7460a91f1265f6101e951bc82ac76" exitCode=0 Feb 20 08:11:11 crc kubenswrapper[4948]: I0220 08:11:11.198017 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbtm5" event={"ID":"5ee242c9-9783-44cb-a4dc-1ea50a7ae185","Type":"ContainerDied","Data":"1da87f05a71242abf8441f6ff99031f27ba7460a91f1265f6101e951bc82ac76"} Feb 20 08:11:11 crc kubenswrapper[4948]: I0220 08:11:11.221720 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x7bcm" podStartSLOduration=2.826067473 podStartE2EDuration="4.221699519s" podCreationTimestamp="2026-02-20 08:11:07 +0000 UTC" firstStartedPulling="2026-02-20 08:11:09.165405741 +0000 UTC m=+318.139900591" lastFinishedPulling="2026-02-20 08:11:10.561037817 +0000 UTC m=+319.535532637" observedRunningTime="2026-02-20 08:11:11.219212198 +0000 UTC m=+320.193707018" watchObservedRunningTime="2026-02-20 08:11:11.221699519 +0000 UTC m=+320.196194349" Feb 20 08:11:11 crc kubenswrapper[4948]: I0220 08:11:11.761039 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fhfhr"] Feb 20 08:11:11 crc kubenswrapper[4948]: I0220 08:11:11.761746 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" podUID="21db7edd-6db3-4afa-b470-0016a9c5afb9" containerName="controller-manager" containerID="cri-o://aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e" gracePeriod=30 Feb 20 08:11:11 crc kubenswrapper[4948]: I0220 08:11:11.859434 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk"] Feb 20 08:11:11 crc kubenswrapper[4948]: I0220 08:11:11.859725 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" podUID="6c070b67-c2ca-459b-a1b7-813a8833e27e" containerName="route-controller-manager" containerID="cri-o://27c38be8b9e4ccb4d5b9ff6afe53e0b0eb4b257a4accae7cd8daf9b3703d5ce8" gracePeriod=30 Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.172467 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.188482 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21db7edd-6db3-4afa-b470-0016a9c5afb9-serving-cert\") pod \"21db7edd-6db3-4afa-b470-0016a9c5afb9\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.188610 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-proxy-ca-bundles\") pod \"21db7edd-6db3-4afa-b470-0016a9c5afb9\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.188665 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-config\") pod \"21db7edd-6db3-4afa-b470-0016a9c5afb9\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.188774 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-client-ca\") pod \"21db7edd-6db3-4afa-b470-0016a9c5afb9\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.190335 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsbfs\" (UniqueName: \"kubernetes.io/projected/21db7edd-6db3-4afa-b470-0016a9c5afb9-kube-api-access-vsbfs\") pod \"21db7edd-6db3-4afa-b470-0016a9c5afb9\" (UID: \"21db7edd-6db3-4afa-b470-0016a9c5afb9\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.190430 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "21db7edd-6db3-4afa-b470-0016a9c5afb9" (UID: "21db7edd-6db3-4afa-b470-0016a9c5afb9"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.190816 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.190985 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-config" (OuterVolumeSpecName: "config") pod "21db7edd-6db3-4afa-b470-0016a9c5afb9" (UID: "21db7edd-6db3-4afa-b470-0016a9c5afb9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.191362 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-client-ca" (OuterVolumeSpecName: "client-ca") pod "21db7edd-6db3-4afa-b470-0016a9c5afb9" (UID: "21db7edd-6db3-4afa-b470-0016a9c5afb9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.197475 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21db7edd-6db3-4afa-b470-0016a9c5afb9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "21db7edd-6db3-4afa-b470-0016a9c5afb9" (UID: "21db7edd-6db3-4afa-b470-0016a9c5afb9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.198571 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21db7edd-6db3-4afa-b470-0016a9c5afb9-kube-api-access-vsbfs" (OuterVolumeSpecName: "kube-api-access-vsbfs") pod "21db7edd-6db3-4afa-b470-0016a9c5afb9" (UID: "21db7edd-6db3-4afa-b470-0016a9c5afb9"). InnerVolumeSpecName "kube-api-access-vsbfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.211479 4948 generic.go:334] "Generic (PLEG): container finished" podID="6c070b67-c2ca-459b-a1b7-813a8833e27e" containerID="27c38be8b9e4ccb4d5b9ff6afe53e0b0eb4b257a4accae7cd8daf9b3703d5ce8" exitCode=0 Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.211562 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" event={"ID":"6c070b67-c2ca-459b-a1b7-813a8833e27e","Type":"ContainerDied","Data":"27c38be8b9e4ccb4d5b9ff6afe53e0b0eb4b257a4accae7cd8daf9b3703d5ce8"} Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.213445 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbtm5" event={"ID":"5ee242c9-9783-44cb-a4dc-1ea50a7ae185","Type":"ContainerStarted","Data":"3b29859986cd536efddb95cc16c282cbd1fab4057f4305e4757cce50ace32d1a"} Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.226841 4948 generic.go:334] "Generic (PLEG): container finished" podID="21db7edd-6db3-4afa-b470-0016a9c5afb9" containerID="aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e" exitCode=0 Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.227258 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" event={"ID":"21db7edd-6db3-4afa-b470-0016a9c5afb9","Type":"ContainerDied","Data":"aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e"} Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.227334 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" event={"ID":"21db7edd-6db3-4afa-b470-0016a9c5afb9","Type":"ContainerDied","Data":"84341e8559dd79e093cb4de6142b9a22ac6a50040f68f50623fb6ec98aa5e33a"} Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.227360 4948 scope.go:117] "RemoveContainer" containerID="aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.227210 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fhfhr" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.250090 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jbtm5" podStartSLOduration=1.844780842 podStartE2EDuration="4.250070229s" podCreationTimestamp="2026-02-20 08:11:08 +0000 UTC" firstStartedPulling="2026-02-20 08:11:09.175053347 +0000 UTC m=+318.149548177" lastFinishedPulling="2026-02-20 08:11:11.580342744 +0000 UTC m=+320.554837564" observedRunningTime="2026-02-20 08:11:12.247366012 +0000 UTC m=+321.221860832" watchObservedRunningTime="2026-02-20 08:11:12.250070229 +0000 UTC m=+321.224565049" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.254941 4948 scope.go:117] "RemoveContainer" containerID="aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e" Feb 20 08:11:12 crc kubenswrapper[4948]: E0220 08:11:12.255722 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e\": container with ID starting with aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e not found: ID does not exist" containerID="aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.255757 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e"} err="failed to get container status \"aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e\": rpc error: code = NotFound desc = could not find container \"aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e\": container with ID starting with aa75719650689fb107484882a72b15a095aba1554265c006218efdf71c2e2e3e not found: ID does not exist" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.256757 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.285659 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fhfhr"] Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.292027 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.292053 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21db7edd-6db3-4afa-b470-0016a9c5afb9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.292066 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsbfs\" (UniqueName: \"kubernetes.io/projected/21db7edd-6db3-4afa-b470-0016a9c5afb9-kube-api-access-vsbfs\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.292076 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21db7edd-6db3-4afa-b470-0016a9c5afb9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.299051 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fhfhr"] Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.392419 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-config\") pod \"6c070b67-c2ca-459b-a1b7-813a8833e27e\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.392476 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c070b67-c2ca-459b-a1b7-813a8833e27e-serving-cert\") pod \"6c070b67-c2ca-459b-a1b7-813a8833e27e\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.392516 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-client-ca\") pod \"6c070b67-c2ca-459b-a1b7-813a8833e27e\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.392594 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25nlb\" (UniqueName: \"kubernetes.io/projected/6c070b67-c2ca-459b-a1b7-813a8833e27e-kube-api-access-25nlb\") pod \"6c070b67-c2ca-459b-a1b7-813a8833e27e\" (UID: \"6c070b67-c2ca-459b-a1b7-813a8833e27e\") " Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.393995 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-client-ca" (OuterVolumeSpecName: "client-ca") pod "6c070b67-c2ca-459b-a1b7-813a8833e27e" (UID: "6c070b67-c2ca-459b-a1b7-813a8833e27e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.394185 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-config" (OuterVolumeSpecName: "config") pod "6c070b67-c2ca-459b-a1b7-813a8833e27e" (UID: "6c070b67-c2ca-459b-a1b7-813a8833e27e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.397815 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c070b67-c2ca-459b-a1b7-813a8833e27e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6c070b67-c2ca-459b-a1b7-813a8833e27e" (UID: "6c070b67-c2ca-459b-a1b7-813a8833e27e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.398197 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c070b67-c2ca-459b-a1b7-813a8833e27e-kube-api-access-25nlb" (OuterVolumeSpecName: "kube-api-access-25nlb") pod "6c070b67-c2ca-459b-a1b7-813a8833e27e" (UID: "6c070b67-c2ca-459b-a1b7-813a8833e27e"). InnerVolumeSpecName "kube-api-access-25nlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.493706 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25nlb\" (UniqueName: \"kubernetes.io/projected/6c070b67-c2ca-459b-a1b7-813a8833e27e-kube-api-access-25nlb\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.493744 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.493754 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c070b67-c2ca-459b-a1b7-813a8833e27e-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:12 crc kubenswrapper[4948]: I0220 08:11:12.493762 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c070b67-c2ca-459b-a1b7-813a8833e27e-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.237514 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.237526 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk" event={"ID":"6c070b67-c2ca-459b-a1b7-813a8833e27e","Type":"ContainerDied","Data":"f24acc5a18bb988f4e5c8c67e817d3ea2636025712e6b3477a09950d4595e36b"} Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.237622 4948 scope.go:117] "RemoveContainer" containerID="27c38be8b9e4ccb4d5b9ff6afe53e0b0eb4b257a4accae7cd8daf9b3703d5ce8" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.296504 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk"] Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.302468 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qljmk"] Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.346352 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5dfff44d5b-x9695"] Feb 20 08:11:13 crc kubenswrapper[4948]: E0220 08:11:13.346665 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21db7edd-6db3-4afa-b470-0016a9c5afb9" containerName="controller-manager" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.346687 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="21db7edd-6db3-4afa-b470-0016a9c5afb9" containerName="controller-manager" Feb 20 08:11:13 crc kubenswrapper[4948]: E0220 08:11:13.346729 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c070b67-c2ca-459b-a1b7-813a8833e27e" containerName="route-controller-manager" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.346737 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c070b67-c2ca-459b-a1b7-813a8833e27e" containerName="route-controller-manager" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.346857 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c070b67-c2ca-459b-a1b7-813a8833e27e" containerName="route-controller-manager" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.346881 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="21db7edd-6db3-4afa-b470-0016a9c5afb9" containerName="controller-manager" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.347401 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.350250 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.351077 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.351653 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.354800 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.355031 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.356852 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.360434 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4"] Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.361261 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.363926 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.364002 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.364469 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.364998 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.365369 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.365833 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.368324 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.369935 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5dfff44d5b-x9695"] Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.376523 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4"] Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.506921 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-config\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.507033 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-proxy-ca-bundles\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.507113 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-client-ca\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.507229 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb83d65b-730e-4024-a9a8-27b2a18b305d-serving-cert\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.507273 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-serving-cert\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.507327 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-config\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.507442 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljpsl\" (UniqueName: \"kubernetes.io/projected/fb83d65b-730e-4024-a9a8-27b2a18b305d-kube-api-access-ljpsl\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.507480 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jng5g\" (UniqueName: \"kubernetes.io/projected/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-kube-api-access-jng5g\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.507529 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-client-ca\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608415 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-config\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608508 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-proxy-ca-bundles\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608566 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-client-ca\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608621 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb83d65b-730e-4024-a9a8-27b2a18b305d-serving-cert\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608660 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-serving-cert\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608701 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-config\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608758 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljpsl\" (UniqueName: \"kubernetes.io/projected/fb83d65b-730e-4024-a9a8-27b2a18b305d-kube-api-access-ljpsl\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608792 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jng5g\" (UniqueName: \"kubernetes.io/projected/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-kube-api-access-jng5g\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.608862 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-client-ca\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.609987 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-config\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.610366 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-client-ca\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.610478 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-client-ca\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.610602 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-config\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.611286 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fb83d65b-730e-4024-a9a8-27b2a18b305d-proxy-ca-bundles\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.614502 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-serving-cert\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.614692 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb83d65b-730e-4024-a9a8-27b2a18b305d-serving-cert\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.631706 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jng5g\" (UniqueName: \"kubernetes.io/projected/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-kube-api-access-jng5g\") pod \"route-controller-manager-c99b4965f-dm2z4\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.642514 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljpsl\" (UniqueName: \"kubernetes.io/projected/fb83d65b-730e-4024-a9a8-27b2a18b305d-kube-api-access-ljpsl\") pod \"controller-manager-5dfff44d5b-x9695\" (UID: \"fb83d65b-730e-4024-a9a8-27b2a18b305d\") " pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.683440 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.700084 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.729416 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21db7edd-6db3-4afa-b470-0016a9c5afb9" path="/var/lib/kubelet/pods/21db7edd-6db3-4afa-b470-0016a9c5afb9/volumes" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.730400 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c070b67-c2ca-459b-a1b7-813a8833e27e" path="/var/lib/kubelet/pods/6c070b67-c2ca-459b-a1b7-813a8833e27e/volumes" Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.952777 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4"] Feb 20 08:11:13 crc kubenswrapper[4948]: W0220 08:11:13.962227 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fba2cd6_ab12_40bb_90e7_17681ddc5bcf.slice/crio-b7b0531bcc9a09086ceb7e7ef51d7fe67486fce64c4091ce69022bf2050f95fe WatchSource:0}: Error finding container b7b0531bcc9a09086ceb7e7ef51d7fe67486fce64c4091ce69022bf2050f95fe: Status 404 returned error can't find the container with id b7b0531bcc9a09086ceb7e7ef51d7fe67486fce64c4091ce69022bf2050f95fe Feb 20 08:11:13 crc kubenswrapper[4948]: I0220 08:11:13.990726 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5dfff44d5b-x9695"] Feb 20 08:11:13 crc kubenswrapper[4948]: W0220 08:11:13.998101 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb83d65b_730e_4024_a9a8_27b2a18b305d.slice/crio-3390c540d19ba441062645bacaa2af7d0d4366dcacd7de98b84bc73030b8affd WatchSource:0}: Error finding container 3390c540d19ba441062645bacaa2af7d0d4366dcacd7de98b84bc73030b8affd: Status 404 returned error can't find the container with id 3390c540d19ba441062645bacaa2af7d0d4366dcacd7de98b84bc73030b8affd Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.245944 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" event={"ID":"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf","Type":"ContainerStarted","Data":"e01ead6dca19e54015aea007496c57209ea14c70dd72f890df0bf653660cc282"} Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.246016 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" event={"ID":"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf","Type":"ContainerStarted","Data":"b7b0531bcc9a09086ceb7e7ef51d7fe67486fce64c4091ce69022bf2050f95fe"} Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.246490 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.247617 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" event={"ID":"fb83d65b-730e-4024-a9a8-27b2a18b305d","Type":"ContainerStarted","Data":"ccba885eaf6bc7f27fba8d3843893c2538dc7e82d238fe2e090637adf95f2666"} Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.247648 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" event={"ID":"fb83d65b-730e-4024-a9a8-27b2a18b305d","Type":"ContainerStarted","Data":"3390c540d19ba441062645bacaa2af7d0d4366dcacd7de98b84bc73030b8affd"} Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.248118 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.252830 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.262531 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" podStartSLOduration=3.26251091 podStartE2EDuration="3.26251091s" podCreationTimestamp="2026-02-20 08:11:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:11:14.2604499 +0000 UTC m=+323.234944760" watchObservedRunningTime="2026-02-20 08:11:14.26251091 +0000 UTC m=+323.237005730" Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.492838 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:14 crc kubenswrapper[4948]: I0220 08:11:14.522515 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5dfff44d5b-x9695" podStartSLOduration=3.522500976 podStartE2EDuration="3.522500976s" podCreationTimestamp="2026-02-20 08:11:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:11:14.281374703 +0000 UTC m=+323.255869543" watchObservedRunningTime="2026-02-20 08:11:14.522500976 +0000 UTC m=+323.496995796" Feb 20 08:11:15 crc kubenswrapper[4948]: I0220 08:11:15.382034 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4"] Feb 20 08:11:15 crc kubenswrapper[4948]: I0220 08:11:15.845204 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:15 crc kubenswrapper[4948]: I0220 08:11:15.845284 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:15 crc kubenswrapper[4948]: I0220 08:11:15.906289 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:15 crc kubenswrapper[4948]: I0220 08:11:15.991173 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:15 crc kubenswrapper[4948]: I0220 08:11:15.991532 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:16 crc kubenswrapper[4948]: I0220 08:11:16.039740 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:16 crc kubenswrapper[4948]: I0220 08:11:16.308209 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tt6t6" Feb 20 08:11:16 crc kubenswrapper[4948]: I0220 08:11:16.319307 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dl7nj" Feb 20 08:11:17 crc kubenswrapper[4948]: I0220 08:11:17.268146 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" podUID="2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" containerName="route-controller-manager" containerID="cri-o://e01ead6dca19e54015aea007496c57209ea14c70dd72f890df0bf653660cc282" gracePeriod=30 Feb 20 08:11:18 crc kubenswrapper[4948]: I0220 08:11:18.216966 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:18 crc kubenswrapper[4948]: I0220 08:11:18.217041 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:18 crc kubenswrapper[4948]: I0220 08:11:18.262687 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:18 crc kubenswrapper[4948]: I0220 08:11:18.344800 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x7bcm" Feb 20 08:11:18 crc kubenswrapper[4948]: I0220 08:11:18.467626 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:18 crc kubenswrapper[4948]: I0220 08:11:18.467682 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.281397 4948 generic.go:334] "Generic (PLEG): container finished" podID="2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" containerID="e01ead6dca19e54015aea007496c57209ea14c70dd72f890df0bf653660cc282" exitCode=0 Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.281540 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" event={"ID":"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf","Type":"ContainerDied","Data":"e01ead6dca19e54015aea007496c57209ea14c70dd72f890df0bf653660cc282"} Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.501081 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jbtm5" podUID="5ee242c9-9783-44cb-a4dc-1ea50a7ae185" containerName="registry-server" probeResult="failure" output=< Feb 20 08:11:19 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 08:11:19 crc kubenswrapper[4948]: > Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.689054 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.717652 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m"] Feb 20 08:11:19 crc kubenswrapper[4948]: E0220 08:11:19.717998 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" containerName="route-controller-manager" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.718020 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" containerName="route-controller-manager" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.718257 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" containerName="route-controller-manager" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.718839 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.736849 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m"] Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.859846 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jng5g\" (UniqueName: \"kubernetes.io/projected/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-kube-api-access-jng5g\") pod \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.860340 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-client-ca\") pod \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.860512 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-serving-cert\") pod \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.860680 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-config\") pod \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\" (UID: \"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf\") " Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.860930 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8p8l\" (UniqueName: \"kubernetes.io/projected/7d1617e1-d79c-4197-ac74-edc70e60870d-kube-api-access-z8p8l\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.861061 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d1617e1-d79c-4197-ac74-edc70e60870d-serving-cert\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.861217 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-client-ca\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.861226 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-client-ca" (OuterVolumeSpecName: "client-ca") pod "2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" (UID: "2fba2cd6-ab12-40bb-90e7-17681ddc5bcf"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.861365 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-config" (OuterVolumeSpecName: "config") pod "2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" (UID: "2fba2cd6-ab12-40bb-90e7-17681ddc5bcf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.861684 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-config\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.861870 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.861892 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.869246 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-kube-api-access-jng5g" (OuterVolumeSpecName: "kube-api-access-jng5g") pod "2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" (UID: "2fba2cd6-ab12-40bb-90e7-17681ddc5bcf"). InnerVolumeSpecName "kube-api-access-jng5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.869508 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" (UID: "2fba2cd6-ab12-40bb-90e7-17681ddc5bcf"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.963041 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8p8l\" (UniqueName: \"kubernetes.io/projected/7d1617e1-d79c-4197-ac74-edc70e60870d-kube-api-access-z8p8l\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.963092 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d1617e1-d79c-4197-ac74-edc70e60870d-serving-cert\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.963131 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-client-ca\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.963151 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-config\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.963186 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jng5g\" (UniqueName: \"kubernetes.io/projected/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-kube-api-access-jng5g\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.963201 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.964324 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-client-ca\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.964524 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-config\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.966953 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d1617e1-d79c-4197-ac74-edc70e60870d-serving-cert\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:19 crc kubenswrapper[4948]: I0220 08:11:19.982018 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8p8l\" (UniqueName: \"kubernetes.io/projected/7d1617e1-d79c-4197-ac74-edc70e60870d-kube-api-access-z8p8l\") pod \"route-controller-manager-7565cf79bc-8x44m\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:20 crc kubenswrapper[4948]: I0220 08:11:20.054440 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:20 crc kubenswrapper[4948]: I0220 08:11:20.288638 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" Feb 20 08:11:20 crc kubenswrapper[4948]: I0220 08:11:20.291381 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4" event={"ID":"2fba2cd6-ab12-40bb-90e7-17681ddc5bcf","Type":"ContainerDied","Data":"b7b0531bcc9a09086ceb7e7ef51d7fe67486fce64c4091ce69022bf2050f95fe"} Feb 20 08:11:20 crc kubenswrapper[4948]: I0220 08:11:20.291484 4948 scope.go:117] "RemoveContainer" containerID="e01ead6dca19e54015aea007496c57209ea14c70dd72f890df0bf653660cc282" Feb 20 08:11:20 crc kubenswrapper[4948]: I0220 08:11:20.317735 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4"] Feb 20 08:11:20 crc kubenswrapper[4948]: I0220 08:11:20.321208 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-dm2z4"] Feb 20 08:11:20 crc kubenswrapper[4948]: I0220 08:11:20.472157 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m"] Feb 20 08:11:21 crc kubenswrapper[4948]: I0220 08:11:21.295775 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" event={"ID":"7d1617e1-d79c-4197-ac74-edc70e60870d","Type":"ContainerStarted","Data":"64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36"} Feb 20 08:11:21 crc kubenswrapper[4948]: I0220 08:11:21.296166 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" event={"ID":"7d1617e1-d79c-4197-ac74-edc70e60870d","Type":"ContainerStarted","Data":"946f59be51177e033bd3a18511a444f58bf8d8e903a9bda2540b30b85ad59a9f"} Feb 20 08:11:21 crc kubenswrapper[4948]: I0220 08:11:21.296470 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:21 crc kubenswrapper[4948]: I0220 08:11:21.304789 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:21 crc kubenswrapper[4948]: I0220 08:11:21.321774 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" podStartSLOduration=6.321758095 podStartE2EDuration="6.321758095s" podCreationTimestamp="2026-02-20 08:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:11:21.321257632 +0000 UTC m=+330.295752462" watchObservedRunningTime="2026-02-20 08:11:21.321758095 +0000 UTC m=+330.296252935" Feb 20 08:11:21 crc kubenswrapper[4948]: I0220 08:11:21.727869 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fba2cd6-ab12-40bb-90e7-17681ddc5bcf" path="/var/lib/kubelet/pods/2fba2cd6-ab12-40bb-90e7-17681ddc5bcf/volumes" Feb 20 08:11:28 crc kubenswrapper[4948]: I0220 08:11:28.537164 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:28 crc kubenswrapper[4948]: I0220 08:11:28.599875 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jbtm5" Feb 20 08:11:38 crc kubenswrapper[4948]: I0220 08:11:38.025441 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:11:38 crc kubenswrapper[4948]: I0220 08:11:38.026733 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.879894 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hhhfd"] Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.882122 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.905318 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hhhfd"] Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.987120 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d804474b-3915-46f1-9fc7-366ccfb6602b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.987175 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-registry-tls\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.987212 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.987242 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d804474b-3915-46f1-9fc7-366ccfb6602b-trusted-ca\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.987340 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d804474b-3915-46f1-9fc7-366ccfb6602b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.987364 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-bound-sa-token\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.987394 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d804474b-3915-46f1-9fc7-366ccfb6602b-registry-certificates\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:46 crc kubenswrapper[4948]: I0220 08:11:46.987431 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbnlk\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-kube-api-access-hbnlk\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.011450 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.089154 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d804474b-3915-46f1-9fc7-366ccfb6602b-registry-certificates\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.089225 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbnlk\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-kube-api-access-hbnlk\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.089291 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d804474b-3915-46f1-9fc7-366ccfb6602b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.089320 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-registry-tls\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.089349 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d804474b-3915-46f1-9fc7-366ccfb6602b-trusted-ca\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.089393 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d804474b-3915-46f1-9fc7-366ccfb6602b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.089414 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-bound-sa-token\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.091056 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d804474b-3915-46f1-9fc7-366ccfb6602b-registry-certificates\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.095353 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d804474b-3915-46f1-9fc7-366ccfb6602b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.112765 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d804474b-3915-46f1-9fc7-366ccfb6602b-trusted-ca\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.117347 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-registry-tls\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.119255 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbnlk\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-kube-api-access-hbnlk\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.121572 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d804474b-3915-46f1-9fc7-366ccfb6602b-bound-sa-token\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.121631 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d804474b-3915-46f1-9fc7-366ccfb6602b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hhhfd\" (UID: \"d804474b-3915-46f1-9fc7-366ccfb6602b\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.204514 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:47 crc kubenswrapper[4948]: I0220 08:11:47.660733 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hhhfd"] Feb 20 08:11:48 crc kubenswrapper[4948]: I0220 08:11:48.491171 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" event={"ID":"d804474b-3915-46f1-9fc7-366ccfb6602b","Type":"ContainerStarted","Data":"96395697398b7e725b35e6e3db0a9fecbe5007e3babb791ce6d7d562c98144b0"} Feb 20 08:11:48 crc kubenswrapper[4948]: I0220 08:11:48.491655 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:11:48 crc kubenswrapper[4948]: I0220 08:11:48.491672 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" event={"ID":"d804474b-3915-46f1-9fc7-366ccfb6602b","Type":"ContainerStarted","Data":"bc2abfbb23d88c80d5df543d5cf4ecd494a9e31b5c733d2e7c759003511f1ec0"} Feb 20 08:11:48 crc kubenswrapper[4948]: I0220 08:11:48.520359 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" podStartSLOduration=2.520324075 podStartE2EDuration="2.520324075s" podCreationTimestamp="2026-02-20 08:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:11:48.520023727 +0000 UTC m=+357.494518547" watchObservedRunningTime="2026-02-20 08:11:48.520324075 +0000 UTC m=+357.494818935" Feb 20 08:11:51 crc kubenswrapper[4948]: I0220 08:11:51.756063 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m"] Feb 20 08:11:51 crc kubenswrapper[4948]: I0220 08:11:51.757745 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" podUID="7d1617e1-d79c-4197-ac74-edc70e60870d" containerName="route-controller-manager" containerID="cri-o://64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36" gracePeriod=30 Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.212514 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.279527 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d1617e1-d79c-4197-ac74-edc70e60870d-serving-cert\") pod \"7d1617e1-d79c-4197-ac74-edc70e60870d\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.279610 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-config\") pod \"7d1617e1-d79c-4197-ac74-edc70e60870d\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.279669 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8p8l\" (UniqueName: \"kubernetes.io/projected/7d1617e1-d79c-4197-ac74-edc70e60870d-kube-api-access-z8p8l\") pod \"7d1617e1-d79c-4197-ac74-edc70e60870d\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.279785 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-client-ca\") pod \"7d1617e1-d79c-4197-ac74-edc70e60870d\" (UID: \"7d1617e1-d79c-4197-ac74-edc70e60870d\") " Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.280256 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-config" (OuterVolumeSpecName: "config") pod "7d1617e1-d79c-4197-ac74-edc70e60870d" (UID: "7d1617e1-d79c-4197-ac74-edc70e60870d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.280398 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-client-ca" (OuterVolumeSpecName: "client-ca") pod "7d1617e1-d79c-4197-ac74-edc70e60870d" (UID: "7d1617e1-d79c-4197-ac74-edc70e60870d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.285876 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d1617e1-d79c-4197-ac74-edc70e60870d-kube-api-access-z8p8l" (OuterVolumeSpecName: "kube-api-access-z8p8l") pod "7d1617e1-d79c-4197-ac74-edc70e60870d" (UID: "7d1617e1-d79c-4197-ac74-edc70e60870d"). InnerVolumeSpecName "kube-api-access-z8p8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.288921 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d1617e1-d79c-4197-ac74-edc70e60870d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7d1617e1-d79c-4197-ac74-edc70e60870d" (UID: "7d1617e1-d79c-4197-ac74-edc70e60870d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.381846 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.381887 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d1617e1-d79c-4197-ac74-edc70e60870d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.381902 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1617e1-d79c-4197-ac74-edc70e60870d-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.381914 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8p8l\" (UniqueName: \"kubernetes.io/projected/7d1617e1-d79c-4197-ac74-edc70e60870d-kube-api-access-z8p8l\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.517438 4948 generic.go:334] "Generic (PLEG): container finished" podID="7d1617e1-d79c-4197-ac74-edc70e60870d" containerID="64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36" exitCode=0 Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.517488 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" event={"ID":"7d1617e1-d79c-4197-ac74-edc70e60870d","Type":"ContainerDied","Data":"64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36"} Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.517515 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" event={"ID":"7d1617e1-d79c-4197-ac74-edc70e60870d","Type":"ContainerDied","Data":"946f59be51177e033bd3a18511a444f58bf8d8e903a9bda2540b30b85ad59a9f"} Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.517542 4948 scope.go:117] "RemoveContainer" containerID="64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.517544 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.535890 4948 scope.go:117] "RemoveContainer" containerID="64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36" Feb 20 08:11:52 crc kubenswrapper[4948]: E0220 08:11:52.536580 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36\": container with ID starting with 64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36 not found: ID does not exist" containerID="64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.536645 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36"} err="failed to get container status \"64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36\": rpc error: code = NotFound desc = could not find container \"64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36\": container with ID starting with 64ad04f585e964b7677390442b789797ba61c7d5f4d2e507fba783bd201bcc36 not found: ID does not exist" Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.551790 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m"] Feb 20 08:11:52 crc kubenswrapper[4948]: I0220 08:11:52.559939 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7565cf79bc-8x44m"] Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.368316 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9"] Feb 20 08:11:53 crc kubenswrapper[4948]: E0220 08:11:53.368878 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d1617e1-d79c-4197-ac74-edc70e60870d" containerName="route-controller-manager" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.368896 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d1617e1-d79c-4197-ac74-edc70e60870d" containerName="route-controller-manager" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.369028 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d1617e1-d79c-4197-ac74-edc70e60870d" containerName="route-controller-manager" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.369479 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.371958 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.372834 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.373208 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.373411 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.373576 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.374023 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.386469 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9"] Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.496146 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05df7d38-ed40-42b4-9761-0ff40846f561-config\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.496222 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05df7d38-ed40-42b4-9761-0ff40846f561-serving-cert\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.496243 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spshl\" (UniqueName: \"kubernetes.io/projected/05df7d38-ed40-42b4-9761-0ff40846f561-kube-api-access-spshl\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.496277 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05df7d38-ed40-42b4-9761-0ff40846f561-client-ca\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.597928 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05df7d38-ed40-42b4-9761-0ff40846f561-config\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.598270 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05df7d38-ed40-42b4-9761-0ff40846f561-serving-cert\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.598339 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spshl\" (UniqueName: \"kubernetes.io/projected/05df7d38-ed40-42b4-9761-0ff40846f561-kube-api-access-spshl\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.598443 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05df7d38-ed40-42b4-9761-0ff40846f561-client-ca\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.599265 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05df7d38-ed40-42b4-9761-0ff40846f561-client-ca\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.599500 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05df7d38-ed40-42b4-9761-0ff40846f561-config\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.603202 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05df7d38-ed40-42b4-9761-0ff40846f561-serving-cert\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.627043 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spshl\" (UniqueName: \"kubernetes.io/projected/05df7d38-ed40-42b4-9761-0ff40846f561-kube-api-access-spshl\") pod \"route-controller-manager-c99b4965f-q7mv9\" (UID: \"05df7d38-ed40-42b4-9761-0ff40846f561\") " pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.685643 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.730751 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d1617e1-d79c-4197-ac74-edc70e60870d" path="/var/lib/kubelet/pods/7d1617e1-d79c-4197-ac74-edc70e60870d/volumes" Feb 20 08:11:53 crc kubenswrapper[4948]: I0220 08:11:53.933072 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9"] Feb 20 08:11:54 crc kubenswrapper[4948]: I0220 08:11:54.530692 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" event={"ID":"05df7d38-ed40-42b4-9761-0ff40846f561","Type":"ContainerStarted","Data":"4af19cbc8b5de82798fcfe4b39a57d6fcf401ac16b4e5756e3169eb3564d0f6f"} Feb 20 08:11:54 crc kubenswrapper[4948]: I0220 08:11:54.531114 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" event={"ID":"05df7d38-ed40-42b4-9761-0ff40846f561","Type":"ContainerStarted","Data":"423ce1385df7d56c79c46fdb192204a1248d92c9c0800a2af62972f7078e6f78"} Feb 20 08:11:54 crc kubenswrapper[4948]: I0220 08:11:54.531348 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:11:54 crc kubenswrapper[4948]: I0220 08:11:54.569170 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" podStartSLOduration=3.569140447 podStartE2EDuration="3.569140447s" podCreationTimestamp="2026-02-20 08:11:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:11:54.565155411 +0000 UTC m=+363.539650231" watchObservedRunningTime="2026-02-20 08:11:54.569140447 +0000 UTC m=+363.543635337" Feb 20 08:11:54 crc kubenswrapper[4948]: I0220 08:11:54.581616 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-c99b4965f-q7mv9" Feb 20 08:12:07 crc kubenswrapper[4948]: I0220 08:12:07.215812 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-hhhfd" Feb 20 08:12:07 crc kubenswrapper[4948]: I0220 08:12:07.281185 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7wm8c"] Feb 20 08:12:08 crc kubenswrapper[4948]: I0220 08:12:08.025194 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:12:08 crc kubenswrapper[4948]: I0220 08:12:08.025300 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.337303 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" podUID="57c455f7-aa9c-405c-bee3-89726b84f5db" containerName="registry" containerID="cri-o://20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120" gracePeriod=30 Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.789932 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.790388 4948 generic.go:334] "Generic (PLEG): container finished" podID="57c455f7-aa9c-405c-bee3-89726b84f5db" containerID="20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120" exitCode=0 Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.790437 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" event={"ID":"57c455f7-aa9c-405c-bee3-89726b84f5db","Type":"ContainerDied","Data":"20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120"} Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.790494 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" event={"ID":"57c455f7-aa9c-405c-bee3-89726b84f5db","Type":"ContainerDied","Data":"f90694ba31dc9edaf32c68d5b113799c7c3cf39817bac7a06e98cc5e157c4d16"} Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.790517 4948 scope.go:117] "RemoveContainer" containerID="20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.804599 4948 scope.go:117] "RemoveContainer" containerID="20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120" Feb 20 08:12:32 crc kubenswrapper[4948]: E0220 08:12:32.805555 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120\": container with ID starting with 20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120 not found: ID does not exist" containerID="20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.805625 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120"} err="failed to get container status \"20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120\": rpc error: code = NotFound desc = could not find container \"20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120\": container with ID starting with 20908318b48f1c9de151f5bb5f43a14fb60c5d750ea49471a9d574dabed15120 not found: ID does not exist" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.943353 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/57c455f7-aa9c-405c-bee3-89726b84f5db-installation-pull-secrets\") pod \"57c455f7-aa9c-405c-bee3-89726b84f5db\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.944140 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-bound-sa-token\") pod \"57c455f7-aa9c-405c-bee3-89726b84f5db\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.944211 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-trusted-ca\") pod \"57c455f7-aa9c-405c-bee3-89726b84f5db\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.944520 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"57c455f7-aa9c-405c-bee3-89726b84f5db\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.944562 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-tls\") pod \"57c455f7-aa9c-405c-bee3-89726b84f5db\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.944608 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-certificates\") pod \"57c455f7-aa9c-405c-bee3-89726b84f5db\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.945253 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vg4v\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-kube-api-access-2vg4v\") pod \"57c455f7-aa9c-405c-bee3-89726b84f5db\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.945357 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/57c455f7-aa9c-405c-bee3-89726b84f5db-ca-trust-extracted\") pod \"57c455f7-aa9c-405c-bee3-89726b84f5db\" (UID: \"57c455f7-aa9c-405c-bee3-89726b84f5db\") " Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.946246 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "57c455f7-aa9c-405c-bee3-89726b84f5db" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.946450 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "57c455f7-aa9c-405c-bee3-89726b84f5db" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.956631 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-kube-api-access-2vg4v" (OuterVolumeSpecName: "kube-api-access-2vg4v") pod "57c455f7-aa9c-405c-bee3-89726b84f5db" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db"). InnerVolumeSpecName "kube-api-access-2vg4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.957359 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57c455f7-aa9c-405c-bee3-89726b84f5db-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "57c455f7-aa9c-405c-bee3-89726b84f5db" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.957888 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "57c455f7-aa9c-405c-bee3-89726b84f5db" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.957991 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "57c455f7-aa9c-405c-bee3-89726b84f5db" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.958335 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "57c455f7-aa9c-405c-bee3-89726b84f5db" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:12:32 crc kubenswrapper[4948]: I0220 08:12:32.983810 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57c455f7-aa9c-405c-bee3-89726b84f5db-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "57c455f7-aa9c-405c-bee3-89726b84f5db" (UID: "57c455f7-aa9c-405c-bee3-89726b84f5db"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.047128 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.047195 4948 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.047224 4948 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/57c455f7-aa9c-405c-bee3-89726b84f5db-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.047251 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vg4v\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-kube-api-access-2vg4v\") on node \"crc\" DevicePath \"\"" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.047274 4948 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/57c455f7-aa9c-405c-bee3-89726b84f5db-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.047298 4948 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/57c455f7-aa9c-405c-bee3-89726b84f5db-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.047321 4948 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/57c455f7-aa9c-405c-bee3-89726b84f5db-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.797725 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.822381 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7wm8c"] Feb 20 08:12:33 crc kubenswrapper[4948]: I0220 08:12:33.826664 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7wm8c"] Feb 20 08:12:35 crc kubenswrapper[4948]: I0220 08:12:35.735478 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57c455f7-aa9c-405c-bee3-89726b84f5db" path="/var/lib/kubelet/pods/57c455f7-aa9c-405c-bee3-89726b84f5db/volumes" Feb 20 08:12:37 crc kubenswrapper[4948]: I0220 08:12:37.597307 4948 patch_prober.go:28] interesting pod/image-registry-697d97f7c8-7wm8c container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.20:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 20 08:12:37 crc kubenswrapper[4948]: I0220 08:12:37.598229 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-697d97f7c8-7wm8c" podUID="57c455f7-aa9c-405c-bee3-89726b84f5db" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.20:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.024531 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.024656 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.024717 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.025561 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6791802a782456e3c45df951fa55097ce0ea3affd6a9878fc2981511046d1b09"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.025705 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://6791802a782456e3c45df951fa55097ce0ea3affd6a9878fc2981511046d1b09" gracePeriod=600 Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.834144 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="6791802a782456e3c45df951fa55097ce0ea3affd6a9878fc2981511046d1b09" exitCode=0 Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.834200 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"6791802a782456e3c45df951fa55097ce0ea3affd6a9878fc2981511046d1b09"} Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.834619 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"a2c247070dfb4d0488751bb46e5543e834b3853c983d98f00e419c6aabfdeb7c"} Feb 20 08:12:38 crc kubenswrapper[4948]: I0220 08:12:38.834642 4948 scope.go:117] "RemoveContainer" containerID="874edde981ff7897b7f895d9594e83d7d1c500024bc1d8fa5d651b00d8d1bd59" Feb 20 08:14:38 crc kubenswrapper[4948]: I0220 08:14:38.024735 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:14:38 crc kubenswrapper[4948]: I0220 08:14:38.025344 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.203298 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9"] Feb 20 08:15:00 crc kubenswrapper[4948]: E0220 08:15:00.204510 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57c455f7-aa9c-405c-bee3-89726b84f5db" containerName="registry" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.204540 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="57c455f7-aa9c-405c-bee3-89726b84f5db" containerName="registry" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.204779 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="57c455f7-aa9c-405c-bee3-89726b84f5db" containerName="registry" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.205623 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.211357 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.212172 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.218276 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9"] Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.384581 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9db735d8-ae70-471b-9d86-3be100901c72-secret-volume\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.384754 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqj9z\" (UniqueName: \"kubernetes.io/projected/9db735d8-ae70-471b-9d86-3be100901c72-kube-api-access-fqj9z\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.384865 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9db735d8-ae70-471b-9d86-3be100901c72-config-volume\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.485661 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9db735d8-ae70-471b-9d86-3be100901c72-secret-volume\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.485738 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqj9z\" (UniqueName: \"kubernetes.io/projected/9db735d8-ae70-471b-9d86-3be100901c72-kube-api-access-fqj9z\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.485774 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9db735d8-ae70-471b-9d86-3be100901c72-config-volume\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.486740 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9db735d8-ae70-471b-9d86-3be100901c72-config-volume\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.495368 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9db735d8-ae70-471b-9d86-3be100901c72-secret-volume\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.508321 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqj9z\" (UniqueName: \"kubernetes.io/projected/9db735d8-ae70-471b-9d86-3be100901c72-kube-api-access-fqj9z\") pod \"collect-profiles-29526255-62px9\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.538318 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:00 crc kubenswrapper[4948]: I0220 08:15:00.741669 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9"] Feb 20 08:15:00 crc kubenswrapper[4948]: W0220 08:15:00.752081 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9db735d8_ae70_471b_9d86_3be100901c72.slice/crio-0801fcdfdc4f8abf5fd4ba611ffdd02a9bfe2e69c820029184f902432874be1d WatchSource:0}: Error finding container 0801fcdfdc4f8abf5fd4ba611ffdd02a9bfe2e69c820029184f902432874be1d: Status 404 returned error can't find the container with id 0801fcdfdc4f8abf5fd4ba611ffdd02a9bfe2e69c820029184f902432874be1d Feb 20 08:15:01 crc kubenswrapper[4948]: I0220 08:15:01.332478 4948 generic.go:334] "Generic (PLEG): container finished" podID="9db735d8-ae70-471b-9d86-3be100901c72" containerID="984bcec5b88d4682af52feec68d576f83622efc6d9578f02504c68786eff5cfd" exitCode=0 Feb 20 08:15:01 crc kubenswrapper[4948]: I0220 08:15:01.332733 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" event={"ID":"9db735d8-ae70-471b-9d86-3be100901c72","Type":"ContainerDied","Data":"984bcec5b88d4682af52feec68d576f83622efc6d9578f02504c68786eff5cfd"} Feb 20 08:15:01 crc kubenswrapper[4948]: I0220 08:15:01.333015 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" event={"ID":"9db735d8-ae70-471b-9d86-3be100901c72","Type":"ContainerStarted","Data":"0801fcdfdc4f8abf5fd4ba611ffdd02a9bfe2e69c820029184f902432874be1d"} Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.606279 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.715962 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9db735d8-ae70-471b-9d86-3be100901c72-secret-volume\") pod \"9db735d8-ae70-471b-9d86-3be100901c72\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.716097 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9db735d8-ae70-471b-9d86-3be100901c72-config-volume\") pod \"9db735d8-ae70-471b-9d86-3be100901c72\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.716173 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqj9z\" (UniqueName: \"kubernetes.io/projected/9db735d8-ae70-471b-9d86-3be100901c72-kube-api-access-fqj9z\") pod \"9db735d8-ae70-471b-9d86-3be100901c72\" (UID: \"9db735d8-ae70-471b-9d86-3be100901c72\") " Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.717182 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9db735d8-ae70-471b-9d86-3be100901c72-config-volume" (OuterVolumeSpecName: "config-volume") pod "9db735d8-ae70-471b-9d86-3be100901c72" (UID: "9db735d8-ae70-471b-9d86-3be100901c72"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.722738 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9db735d8-ae70-471b-9d86-3be100901c72-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9db735d8-ae70-471b-9d86-3be100901c72" (UID: "9db735d8-ae70-471b-9d86-3be100901c72"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.723631 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9db735d8-ae70-471b-9d86-3be100901c72-kube-api-access-fqj9z" (OuterVolumeSpecName: "kube-api-access-fqj9z") pod "9db735d8-ae70-471b-9d86-3be100901c72" (UID: "9db735d8-ae70-471b-9d86-3be100901c72"). InnerVolumeSpecName "kube-api-access-fqj9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.818261 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqj9z\" (UniqueName: \"kubernetes.io/projected/9db735d8-ae70-471b-9d86-3be100901c72-kube-api-access-fqj9z\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.818362 4948 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9db735d8-ae70-471b-9d86-3be100901c72-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:02 crc kubenswrapper[4948]: I0220 08:15:02.818391 4948 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9db735d8-ae70-471b-9d86-3be100901c72-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:03 crc kubenswrapper[4948]: I0220 08:15:03.349623 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" event={"ID":"9db735d8-ae70-471b-9d86-3be100901c72","Type":"ContainerDied","Data":"0801fcdfdc4f8abf5fd4ba611ffdd02a9bfe2e69c820029184f902432874be1d"} Feb 20 08:15:03 crc kubenswrapper[4948]: I0220 08:15:03.350443 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0801fcdfdc4f8abf5fd4ba611ffdd02a9bfe2e69c820029184f902432874be1d" Feb 20 08:15:03 crc kubenswrapper[4948]: I0220 08:15:03.350402 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9" Feb 20 08:15:08 crc kubenswrapper[4948]: I0220 08:15:08.024683 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:15:08 crc kubenswrapper[4948]: I0220 08:15:08.024746 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.381988 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-9llfv"] Feb 20 08:15:24 crc kubenswrapper[4948]: E0220 08:15:24.384754 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9db735d8-ae70-471b-9d86-3be100901c72" containerName="collect-profiles" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.384771 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="9db735d8-ae70-471b-9d86-3be100901c72" containerName="collect-profiles" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.386343 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="9db735d8-ae70-471b-9d86-3be100901c72" containerName="collect-profiles" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.387179 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-9llfv" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.390601 4948 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-k2zw5" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.392185 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.392230 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.392470 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-smtqw"] Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.393525 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-smtqw" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.394927 4948 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-v2wrp" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.401125 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-9llfv"] Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.424237 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-smtqw"] Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.436856 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-f4bjc"] Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.437693 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.441523 4948 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-jqmnm" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.450434 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-f4bjc"] Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.552832 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftvfl\" (UniqueName: \"kubernetes.io/projected/bb56399a-a840-49af-972f-4f32c91efed7-kube-api-access-ftvfl\") pod \"cert-manager-cainjector-cf98fcc89-9llfv\" (UID: \"bb56399a-a840-49af-972f-4f32c91efed7\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-9llfv" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.552973 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8h75\" (UniqueName: \"kubernetes.io/projected/305a17f4-aef3-4036-8fce-3756ff5bbd2f-kube-api-access-l8h75\") pod \"cert-manager-webhook-687f57d79b-f4bjc\" (UID: \"305a17f4-aef3-4036-8fce-3756ff5bbd2f\") " pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.553064 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jrjz\" (UniqueName: \"kubernetes.io/projected/1437d818-abf4-4602-8a28-e88a76e482a9-kube-api-access-9jrjz\") pod \"cert-manager-858654f9db-smtqw\" (UID: \"1437d818-abf4-4602-8a28-e88a76e482a9\") " pod="cert-manager/cert-manager-858654f9db-smtqw" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.654136 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8h75\" (UniqueName: \"kubernetes.io/projected/305a17f4-aef3-4036-8fce-3756ff5bbd2f-kube-api-access-l8h75\") pod \"cert-manager-webhook-687f57d79b-f4bjc\" (UID: \"305a17f4-aef3-4036-8fce-3756ff5bbd2f\") " pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.654238 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jrjz\" (UniqueName: \"kubernetes.io/projected/1437d818-abf4-4602-8a28-e88a76e482a9-kube-api-access-9jrjz\") pod \"cert-manager-858654f9db-smtqw\" (UID: \"1437d818-abf4-4602-8a28-e88a76e482a9\") " pod="cert-manager/cert-manager-858654f9db-smtqw" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.654293 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftvfl\" (UniqueName: \"kubernetes.io/projected/bb56399a-a840-49af-972f-4f32c91efed7-kube-api-access-ftvfl\") pod \"cert-manager-cainjector-cf98fcc89-9llfv\" (UID: \"bb56399a-a840-49af-972f-4f32c91efed7\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-9llfv" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.677920 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8h75\" (UniqueName: \"kubernetes.io/projected/305a17f4-aef3-4036-8fce-3756ff5bbd2f-kube-api-access-l8h75\") pod \"cert-manager-webhook-687f57d79b-f4bjc\" (UID: \"305a17f4-aef3-4036-8fce-3756ff5bbd2f\") " pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.678590 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftvfl\" (UniqueName: \"kubernetes.io/projected/bb56399a-a840-49af-972f-4f32c91efed7-kube-api-access-ftvfl\") pod \"cert-manager-cainjector-cf98fcc89-9llfv\" (UID: \"bb56399a-a840-49af-972f-4f32c91efed7\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-9llfv" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.681000 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jrjz\" (UniqueName: \"kubernetes.io/projected/1437d818-abf4-4602-8a28-e88a76e482a9-kube-api-access-9jrjz\") pod \"cert-manager-858654f9db-smtqw\" (UID: \"1437d818-abf4-4602-8a28-e88a76e482a9\") " pod="cert-manager/cert-manager-858654f9db-smtqw" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.704900 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-9llfv" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.712165 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-smtqw" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.751043 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.983671 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-f4bjc"] Feb 20 08:15:24 crc kubenswrapper[4948]: I0220 08:15:24.999325 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:15:25 crc kubenswrapper[4948]: I0220 08:15:25.132635 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-smtqw"] Feb 20 08:15:25 crc kubenswrapper[4948]: I0220 08:15:25.149846 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-9llfv"] Feb 20 08:15:25 crc kubenswrapper[4948]: W0220 08:15:25.156206 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb56399a_a840_49af_972f_4f32c91efed7.slice/crio-d0eaef1010af91c5fcc7aaadb45551119645a38b302145a9450750c0da2c7c34 WatchSource:0}: Error finding container d0eaef1010af91c5fcc7aaadb45551119645a38b302145a9450750c0da2c7c34: Status 404 returned error can't find the container with id d0eaef1010af91c5fcc7aaadb45551119645a38b302145a9450750c0da2c7c34 Feb 20 08:15:25 crc kubenswrapper[4948]: I0220 08:15:25.501963 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" event={"ID":"305a17f4-aef3-4036-8fce-3756ff5bbd2f","Type":"ContainerStarted","Data":"32bc60676f3e532776c6c21ba8be09d090b22873a836263a5e246756ccba5bb9"} Feb 20 08:15:25 crc kubenswrapper[4948]: I0220 08:15:25.503436 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-smtqw" event={"ID":"1437d818-abf4-4602-8a28-e88a76e482a9","Type":"ContainerStarted","Data":"fbaf331b99a09cd4c0955dfc5844f6e128026c3e9782f085962e44bd1d0ab81b"} Feb 20 08:15:25 crc kubenswrapper[4948]: I0220 08:15:25.505221 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-9llfv" event={"ID":"bb56399a-a840-49af-972f-4f32c91efed7","Type":"ContainerStarted","Data":"d0eaef1010af91c5fcc7aaadb45551119645a38b302145a9450750c0da2c7c34"} Feb 20 08:15:30 crc kubenswrapper[4948]: I0220 08:15:30.542185 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-smtqw" event={"ID":"1437d818-abf4-4602-8a28-e88a76e482a9","Type":"ContainerStarted","Data":"f3c8fec8b780bb055690941032243c45f7653a3e6b93dffb13ac54496d436e7e"} Feb 20 08:15:30 crc kubenswrapper[4948]: I0220 08:15:30.544133 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-9llfv" event={"ID":"bb56399a-a840-49af-972f-4f32c91efed7","Type":"ContainerStarted","Data":"552830168d2299c90df97b8c06eb074c90da6db5087c1bcf8fd36c3079dbacc4"} Feb 20 08:15:30 crc kubenswrapper[4948]: I0220 08:15:30.545984 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" event={"ID":"305a17f4-aef3-4036-8fce-3756ff5bbd2f","Type":"ContainerStarted","Data":"a8d19f39147e5e961003b5aaac53e6b1e65d2c280885386da9cc14e15b28ad8a"} Feb 20 08:15:30 crc kubenswrapper[4948]: I0220 08:15:30.546227 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" Feb 20 08:15:30 crc kubenswrapper[4948]: I0220 08:15:30.570848 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-smtqw" podStartSLOduration=2.39105744 podStartE2EDuration="6.570818521s" podCreationTimestamp="2026-02-20 08:15:24 +0000 UTC" firstStartedPulling="2026-02-20 08:15:25.141489303 +0000 UTC m=+574.115984123" lastFinishedPulling="2026-02-20 08:15:29.321250374 +0000 UTC m=+578.295745204" observedRunningTime="2026-02-20 08:15:30.564042273 +0000 UTC m=+579.538537113" watchObservedRunningTime="2026-02-20 08:15:30.570818521 +0000 UTC m=+579.545313361" Feb 20 08:15:30 crc kubenswrapper[4948]: I0220 08:15:30.581205 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-9llfv" podStartSLOduration=2.4194332960000002 podStartE2EDuration="6.581177289s" podCreationTimestamp="2026-02-20 08:15:24 +0000 UTC" firstStartedPulling="2026-02-20 08:15:25.159793408 +0000 UTC m=+574.134288228" lastFinishedPulling="2026-02-20 08:15:29.321537361 +0000 UTC m=+578.296032221" observedRunningTime="2026-02-20 08:15:30.58044695 +0000 UTC m=+579.554941780" watchObservedRunningTime="2026-02-20 08:15:30.581177289 +0000 UTC m=+579.555672129" Feb 20 08:15:30 crc kubenswrapper[4948]: I0220 08:15:30.618850 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" podStartSLOduration=2.211713786 podStartE2EDuration="6.618823624s" podCreationTimestamp="2026-02-20 08:15:24 +0000 UTC" firstStartedPulling="2026-02-20 08:15:24.999104577 +0000 UTC m=+573.973599397" lastFinishedPulling="2026-02-20 08:15:29.406214405 +0000 UTC m=+578.380709235" observedRunningTime="2026-02-20 08:15:30.61584189 +0000 UTC m=+579.590336750" watchObservedRunningTime="2026-02-20 08:15:30.618823624 +0000 UTC m=+579.593318454" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.622433 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7s8gg"] Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.623772 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovn-controller" containerID="cri-o://9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d" gracePeriod=30 Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.624004 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="northd" containerID="cri-o://5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2" gracePeriod=30 Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.624171 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37" gracePeriod=30 Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.624319 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kube-rbac-proxy-node" containerID="cri-o://719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c" gracePeriod=30 Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.624357 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovn-acl-logging" containerID="cri-o://3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7" gracePeriod=30 Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.624510 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="sbdb" containerID="cri-o://c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022" gracePeriod=30 Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.624928 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="nbdb" containerID="cri-o://8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a" gracePeriod=30 Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.665753 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" containerID="cri-o://682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" gracePeriod=30 Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.754358 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-f4bjc" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.907391 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/3.log" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.910426 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovn-acl-logging/0.log" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.911332 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovn-controller/0.log" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.912068 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960418 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cvhbs"] Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960643 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kube-rbac-proxy-node" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960655 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kube-rbac-proxy-node" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960669 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovn-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960676 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovn-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960684 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovn-acl-logging" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960690 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovn-acl-logging" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960698 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960705 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960712 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kubecfg-setup" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960719 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kubecfg-setup" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960728 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="nbdb" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960734 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="nbdb" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960745 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960753 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960764 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960771 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960781 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="sbdb" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960788 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="sbdb" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960797 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960805 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960814 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kube-rbac-proxy-ovn-metrics" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960821 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kube-rbac-proxy-ovn-metrics" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.960833 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="northd" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960839 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="northd" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960945 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960959 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="sbdb" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.960993 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961004 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kube-rbac-proxy-ovn-metrics" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961013 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovn-acl-logging" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961022 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961030 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961040 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961050 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovn-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961061 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="northd" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961071 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="kube-rbac-proxy-node" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961080 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="nbdb" Feb 20 08:15:34 crc kubenswrapper[4948]: E0220 08:15:34.961368 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.961377 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerName="ovnkube-controller" Feb 20 08:15:34 crc kubenswrapper[4948]: I0220 08:15:34.963213 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.043874 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-config\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044283 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-openvswitch\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044328 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-env-overrides\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044368 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-etc-openvswitch\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044397 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-netns\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044416 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044430 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-systemd-units\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044479 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044491 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-script-lib\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044511 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-ovn\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044531 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044567 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044601 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044704 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-slash\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044726 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-slash" (OuterVolumeSpecName: "host-slash") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044798 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044921 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-node-log\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.044962 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovn-node-metrics-cert\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045006 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045024 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-kubelet\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045036 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045060 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-systemd\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045050 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-node-log" (OuterVolumeSpecName: "node-log") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045099 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-netd\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045106 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045136 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-bin\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045162 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-var-lib-cni-networks-ovn-kubernetes\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045188 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtrs5\" (UniqueName: \"kubernetes.io/projected/d6a28bd7-2b83-43f8-b803-bfe41516e071-kube-api-access-gtrs5\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045211 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-var-lib-openvswitch\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045224 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045229 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-ovn-kubernetes\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045250 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045254 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-log-socket\") pod \"d6a28bd7-2b83-43f8-b803-bfe41516e071\" (UID: \"d6a28bd7-2b83-43f8-b803-bfe41516e071\") " Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045283 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-log-socket" (OuterVolumeSpecName: "log-socket") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045315 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045329 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045365 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-run-netns\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045393 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-cni-bin\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045411 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-cni-netd\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045424 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovnkube-script-lib\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045443 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-kubelet\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045465 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045494 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-systemd-units\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045512 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovnkube-config\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045526 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovn-node-metrics-cert\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045551 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsmb9\" (UniqueName: \"kubernetes.io/projected/0bfc877e-59ed-4803-b07a-b9787b1ab73d-kube-api-access-lsmb9\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045581 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-node-log\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045611 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-systemd\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045628 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-ovn\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045643 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-run-ovn-kubernetes\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045669 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-etc-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045685 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-env-overrides\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045700 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-slash\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045715 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-var-lib-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045745 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-log-socket\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045783 4948 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045793 4948 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045802 4948 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045810 4948 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045818 4948 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045826 4948 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045834 4948 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045842 4948 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-slash\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045852 4948 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-node-log\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045841 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045860 4948 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.045928 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.046066 4948 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.046100 4948 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.046123 4948 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.046142 4948 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-log-socket\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.046159 4948 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.053316 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6a28bd7-2b83-43f8-b803-bfe41516e071-kube-api-access-gtrs5" (OuterVolumeSpecName: "kube-api-access-gtrs5") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "kube-api-access-gtrs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.053322 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.071952 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "d6a28bd7-2b83-43f8-b803-bfe41516e071" (UID: "d6a28bd7-2b83-43f8-b803-bfe41516e071"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.146875 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-etc-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.146940 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-env-overrides\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.146993 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-slash\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147024 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-var-lib-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147070 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-log-socket\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147065 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-etc-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147109 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147173 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147280 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-log-socket\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147232 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-var-lib-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147233 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-run-netns\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147285 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-run-netns\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147443 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-cni-bin\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147181 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-slash\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147500 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-cni-netd\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147550 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-cni-netd\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147560 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-cni-bin\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147572 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovnkube-script-lib\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147660 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-kubelet\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147720 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147743 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-kubelet\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147789 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-systemd-units\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147837 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovnkube-config\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147880 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovn-node-metrics-cert\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147889 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-systemd-units\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147938 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsmb9\" (UniqueName: \"kubernetes.io/projected/0bfc877e-59ed-4803-b07a-b9787b1ab73d-kube-api-access-lsmb9\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.147832 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-openvswitch\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148047 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-node-log\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148146 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-node-log\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148318 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-systemd\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148373 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-ovn\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148413 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-run-ovn-kubernetes\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148461 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-systemd\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148514 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-run-ovn\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148595 4948 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d6a28bd7-2b83-43f8-b803-bfe41516e071-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148601 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0bfc877e-59ed-4803-b07a-b9787b1ab73d-host-run-ovn-kubernetes\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148610 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-env-overrides\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148618 4948 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148708 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtrs5\" (UniqueName: \"kubernetes.io/projected/d6a28bd7-2b83-43f8-b803-bfe41516e071-kube-api-access-gtrs5\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148728 4948 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148747 4948 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d6a28bd7-2b83-43f8-b803-bfe41516e071-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.148774 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovnkube-script-lib\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.149672 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovnkube-config\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.153614 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0bfc877e-59ed-4803-b07a-b9787b1ab73d-ovn-node-metrics-cert\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.176899 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsmb9\" (UniqueName: \"kubernetes.io/projected/0bfc877e-59ed-4803-b07a-b9787b1ab73d-kube-api-access-lsmb9\") pod \"ovnkube-node-cvhbs\" (UID: \"0bfc877e-59ed-4803-b07a-b9787b1ab73d\") " pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.276309 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:35 crc kubenswrapper[4948]: W0220 08:15:35.313065 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0bfc877e_59ed_4803_b07a_b9787b1ab73d.slice/crio-b2bcab2b0e6efcd12fe28db5042d25c8f825151e8a2bf70419baa020ac3e634d WatchSource:0}: Error finding container b2bcab2b0e6efcd12fe28db5042d25c8f825151e8a2bf70419baa020ac3e634d: Status 404 returned error can't find the container with id b2bcab2b0e6efcd12fe28db5042d25c8f825151e8a2bf70419baa020ac3e634d Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.586106 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovnkube-controller/3.log" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.590408 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovn-acl-logging/0.log" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.591284 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7s8gg_d6a28bd7-2b83-43f8-b803-bfe41516e071/ovn-controller/0.log" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592019 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" exitCode=0 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592168 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022" exitCode=0 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592304 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a" exitCode=0 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592417 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2" exitCode=0 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592538 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37" exitCode=0 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592653 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c" exitCode=0 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592754 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7" exitCode=143 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592490 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592895 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6a28bd7-2b83-43f8-b803-bfe41516e071" containerID="9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d" exitCode=143 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.592333 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.593494 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.593666 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.593786 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.593920 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594074 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594217 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594336 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594442 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594539 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594642 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594746 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594857 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.594966 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595101 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595224 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595343 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595460 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595479 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595489 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595502 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595513 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595524 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595534 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595545 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595555 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595572 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595617 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595631 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595642 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595653 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595663 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595674 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595684 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595695 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595705 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595715 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595729 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7s8gg" event={"ID":"d6a28bd7-2b83-43f8-b803-bfe41516e071","Type":"ContainerDied","Data":"0269d0e4618f9bfc08bbd4b396cc89a6bd78be17afe5c9425cf43f02da885bc5"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595745 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595758 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595771 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595781 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595792 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595803 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595813 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595823 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595834 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.595845 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.593839 4948 scope.go:117] "RemoveContainer" containerID="682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.603693 4948 generic.go:334] "Generic (PLEG): container finished" podID="0bfc877e-59ed-4803-b07a-b9787b1ab73d" containerID="83c11a6ecf91713d5e31e2539b1a573dbcda58bd23320b28bcb8d4ab32147fb4" exitCode=0 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.603795 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerDied","Data":"83c11a6ecf91713d5e31e2539b1a573dbcda58bd23320b28bcb8d4ab32147fb4"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.603841 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"b2bcab2b0e6efcd12fe28db5042d25c8f825151e8a2bf70419baa020ac3e634d"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.613690 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/2.log" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.614454 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/1.log" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.614535 4948 generic.go:334] "Generic (PLEG): container finished" podID="b96124e4-0a74-4578-9142-fd728eb9f99e" containerID="4ecf8c7c9ed34410ff4619b18254e2eb90170911300e6cd6a7284e84e8937246" exitCode=2 Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.614581 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7s56" event={"ID":"b96124e4-0a74-4578-9142-fd728eb9f99e","Type":"ContainerDied","Data":"4ecf8c7c9ed34410ff4619b18254e2eb90170911300e6cd6a7284e84e8937246"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.614650 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c"} Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.615432 4948 scope.go:117] "RemoveContainer" containerID="4ecf8c7c9ed34410ff4619b18254e2eb90170911300e6cd6a7284e84e8937246" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.615781 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-x7s56_openshift-multus(b96124e4-0a74-4578-9142-fd728eb9f99e)\"" pod="openshift-multus/multus-x7s56" podUID="b96124e4-0a74-4578-9142-fd728eb9f99e" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.631042 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.691850 4948 scope.go:117] "RemoveContainer" containerID="c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.699640 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7s8gg"] Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.706844 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7s8gg"] Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.719393 4948 scope.go:117] "RemoveContainer" containerID="8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.732012 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6a28bd7-2b83-43f8-b803-bfe41516e071" path="/var/lib/kubelet/pods/d6a28bd7-2b83-43f8-b803-bfe41516e071/volumes" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.743312 4948 scope.go:117] "RemoveContainer" containerID="5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.763432 4948 scope.go:117] "RemoveContainer" containerID="6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.799846 4948 scope.go:117] "RemoveContainer" containerID="719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.814893 4948 scope.go:117] "RemoveContainer" containerID="3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.830619 4948 scope.go:117] "RemoveContainer" containerID="9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.859531 4948 scope.go:117] "RemoveContainer" containerID="540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.893047 4948 scope.go:117] "RemoveContainer" containerID="682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.894633 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": container with ID starting with 682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a not found: ID does not exist" containerID="682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.894689 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} err="failed to get container status \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": rpc error: code = NotFound desc = could not find container \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": container with ID starting with 682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.894727 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.895456 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": container with ID starting with 7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b not found: ID does not exist" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.895508 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} err="failed to get container status \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": rpc error: code = NotFound desc = could not find container \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": container with ID starting with 7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.895542 4948 scope.go:117] "RemoveContainer" containerID="c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.895923 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": container with ID starting with c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022 not found: ID does not exist" containerID="c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.895960 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} err="failed to get container status \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": rpc error: code = NotFound desc = could not find container \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": container with ID starting with c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.896008 4948 scope.go:117] "RemoveContainer" containerID="8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.896863 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": container with ID starting with 8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a not found: ID does not exist" containerID="8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.896948 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} err="failed to get container status \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": rpc error: code = NotFound desc = could not find container \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": container with ID starting with 8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.897099 4948 scope.go:117] "RemoveContainer" containerID="5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.898565 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": container with ID starting with 5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2 not found: ID does not exist" containerID="5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.898714 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} err="failed to get container status \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": rpc error: code = NotFound desc = could not find container \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": container with ID starting with 5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.898755 4948 scope.go:117] "RemoveContainer" containerID="6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.899405 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": container with ID starting with 6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37 not found: ID does not exist" containerID="6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.899444 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} err="failed to get container status \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": rpc error: code = NotFound desc = could not find container \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": container with ID starting with 6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.899472 4948 scope.go:117] "RemoveContainer" containerID="719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.899772 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": container with ID starting with 719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c not found: ID does not exist" containerID="719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.899806 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} err="failed to get container status \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": rpc error: code = NotFound desc = could not find container \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": container with ID starting with 719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.899828 4948 scope.go:117] "RemoveContainer" containerID="3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.900123 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": container with ID starting with 3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7 not found: ID does not exist" containerID="3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.900153 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} err="failed to get container status \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": rpc error: code = NotFound desc = could not find container \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": container with ID starting with 3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.900173 4948 scope.go:117] "RemoveContainer" containerID="9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.900587 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": container with ID starting with 9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d not found: ID does not exist" containerID="9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.900626 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} err="failed to get container status \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": rpc error: code = NotFound desc = could not find container \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": container with ID starting with 9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.900666 4948 scope.go:117] "RemoveContainer" containerID="540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad" Feb 20 08:15:35 crc kubenswrapper[4948]: E0220 08:15:35.900954 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": container with ID starting with 540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad not found: ID does not exist" containerID="540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.901000 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} err="failed to get container status \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": rpc error: code = NotFound desc = could not find container \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": container with ID starting with 540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.901018 4948 scope.go:117] "RemoveContainer" containerID="682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.901305 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} err="failed to get container status \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": rpc error: code = NotFound desc = could not find container \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": container with ID starting with 682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.901331 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.901609 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} err="failed to get container status \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": rpc error: code = NotFound desc = could not find container \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": container with ID starting with 7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.901629 4948 scope.go:117] "RemoveContainer" containerID="c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.901866 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} err="failed to get container status \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": rpc error: code = NotFound desc = could not find container \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": container with ID starting with c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.901894 4948 scope.go:117] "RemoveContainer" containerID="8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.902198 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} err="failed to get container status \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": rpc error: code = NotFound desc = could not find container \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": container with ID starting with 8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.902225 4948 scope.go:117] "RemoveContainer" containerID="5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.902450 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} err="failed to get container status \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": rpc error: code = NotFound desc = could not find container \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": container with ID starting with 5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.902487 4948 scope.go:117] "RemoveContainer" containerID="6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.902685 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} err="failed to get container status \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": rpc error: code = NotFound desc = could not find container \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": container with ID starting with 6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.902717 4948 scope.go:117] "RemoveContainer" containerID="719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.902885 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} err="failed to get container status \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": rpc error: code = NotFound desc = could not find container \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": container with ID starting with 719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.902900 4948 scope.go:117] "RemoveContainer" containerID="3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.903208 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} err="failed to get container status \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": rpc error: code = NotFound desc = could not find container \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": container with ID starting with 3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.903287 4948 scope.go:117] "RemoveContainer" containerID="9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.903550 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} err="failed to get container status \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": rpc error: code = NotFound desc = could not find container \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": container with ID starting with 9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.903635 4948 scope.go:117] "RemoveContainer" containerID="540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.903884 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} err="failed to get container status \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": rpc error: code = NotFound desc = could not find container \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": container with ID starting with 540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.903965 4948 scope.go:117] "RemoveContainer" containerID="682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.904254 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} err="failed to get container status \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": rpc error: code = NotFound desc = could not find container \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": container with ID starting with 682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.904337 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.904648 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} err="failed to get container status \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": rpc error: code = NotFound desc = could not find container \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": container with ID starting with 7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.904678 4948 scope.go:117] "RemoveContainer" containerID="c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.904906 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} err="failed to get container status \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": rpc error: code = NotFound desc = could not find container \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": container with ID starting with c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.905006 4948 scope.go:117] "RemoveContainer" containerID="8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.905446 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} err="failed to get container status \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": rpc error: code = NotFound desc = could not find container \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": container with ID starting with 8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.905507 4948 scope.go:117] "RemoveContainer" containerID="5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.905856 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} err="failed to get container status \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": rpc error: code = NotFound desc = could not find container \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": container with ID starting with 5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.905881 4948 scope.go:117] "RemoveContainer" containerID="6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.906228 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} err="failed to get container status \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": rpc error: code = NotFound desc = could not find container \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": container with ID starting with 6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.906271 4948 scope.go:117] "RemoveContainer" containerID="719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.906579 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} err="failed to get container status \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": rpc error: code = NotFound desc = could not find container \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": container with ID starting with 719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.906605 4948 scope.go:117] "RemoveContainer" containerID="3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.906883 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} err="failed to get container status \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": rpc error: code = NotFound desc = could not find container \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": container with ID starting with 3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.906913 4948 scope.go:117] "RemoveContainer" containerID="9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.907213 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} err="failed to get container status \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": rpc error: code = NotFound desc = could not find container \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": container with ID starting with 9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.907239 4948 scope.go:117] "RemoveContainer" containerID="540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.907453 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} err="failed to get container status \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": rpc error: code = NotFound desc = could not find container \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": container with ID starting with 540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.907539 4948 scope.go:117] "RemoveContainer" containerID="682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.907770 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} err="failed to get container status \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": rpc error: code = NotFound desc = could not find container \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": container with ID starting with 682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.907795 4948 scope.go:117] "RemoveContainer" containerID="7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908046 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b"} err="failed to get container status \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": rpc error: code = NotFound desc = could not find container \"7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b\": container with ID starting with 7c5fabdc5450058d1fdb7d3025e557600f7d0816724e9c2370a40c926260363b not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908072 4948 scope.go:117] "RemoveContainer" containerID="c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908248 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022"} err="failed to get container status \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": rpc error: code = NotFound desc = could not find container \"c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022\": container with ID starting with c152b197ffc0b632f57bf808999e1d725d44b621568ae85c9c032cdc9fb58022 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908273 4948 scope.go:117] "RemoveContainer" containerID="8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908452 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a"} err="failed to get container status \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": rpc error: code = NotFound desc = could not find container \"8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a\": container with ID starting with 8d39bf9ecaf30eaaaf6b775223e06481df334ae2b614b3a042006d2383765a4a not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908475 4948 scope.go:117] "RemoveContainer" containerID="5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908646 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2"} err="failed to get container status \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": rpc error: code = NotFound desc = could not find container \"5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2\": container with ID starting with 5e15d5157065c9f1dc4b4d4c37072145001c540dd6d506182bb20289df5698b2 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908666 4948 scope.go:117] "RemoveContainer" containerID="6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908846 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37"} err="failed to get container status \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": rpc error: code = NotFound desc = could not find container \"6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37\": container with ID starting with 6acdd901fbb1a2c4964cfe350976ca8b303199395e5ffc1152c2f1a05cab0e37 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.908871 4948 scope.go:117] "RemoveContainer" containerID="719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909046 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c"} err="failed to get container status \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": rpc error: code = NotFound desc = could not find container \"719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c\": container with ID starting with 719455d405901c5fbaa4c2e53c6ecc48bcf6ab567c79775a9881d4d32171248c not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909095 4948 scope.go:117] "RemoveContainer" containerID="3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909252 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7"} err="failed to get container status \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": rpc error: code = NotFound desc = could not find container \"3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7\": container with ID starting with 3b51c9f8a2f2d3899a068485a76491a008a382336f3850e321f272d1725731d7 not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909276 4948 scope.go:117] "RemoveContainer" containerID="9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909430 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d"} err="failed to get container status \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": rpc error: code = NotFound desc = could not find container \"9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d\": container with ID starting with 9eb41bd593247d9ce1b71abef38942aa2b54cda5d11569fbf9090c0ffda8474d not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909448 4948 scope.go:117] "RemoveContainer" containerID="540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909606 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad"} err="failed to get container status \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": rpc error: code = NotFound desc = could not find container \"540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad\": container with ID starting with 540f4001c84744bddec572ccbee293aa02b4f1ee7d9dfbdaabbb651996daffad not found: ID does not exist" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909636 4948 scope.go:117] "RemoveContainer" containerID="682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a" Feb 20 08:15:35 crc kubenswrapper[4948]: I0220 08:15:35.909822 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a"} err="failed to get container status \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": rpc error: code = NotFound desc = could not find container \"682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a\": container with ID starting with 682320ab9016f4e642bcf9c863d46d6dc665f3537ab6d443946fc7cb0665d37a not found: ID does not exist" Feb 20 08:15:36 crc kubenswrapper[4948]: I0220 08:15:36.625285 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"fdde3da322ea47cd1001dc4dc1eb54b763197897461c6883cbe223d44b7b759a"} Feb 20 08:15:36 crc kubenswrapper[4948]: I0220 08:15:36.625639 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"b41d6705ce7f06442f224bb66c7e09abe4a15233c1c95a1729006170bae9e8de"} Feb 20 08:15:36 crc kubenswrapper[4948]: I0220 08:15:36.625653 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"a9d7900d2ba119eb35f9472059268560504ebcdb3f51d1d00c88119a2d0ceae9"} Feb 20 08:15:36 crc kubenswrapper[4948]: I0220 08:15:36.625663 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"d8ad6bba55ee80255bca1c489ba329f108b4252e8eec315f332302ef06af7bcd"} Feb 20 08:15:36 crc kubenswrapper[4948]: I0220 08:15:36.625671 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"f52c5394617a409495a9f9b19d57e6549b2a745ab62eaeaf69a3e4f2f9e34f3b"} Feb 20 08:15:36 crc kubenswrapper[4948]: I0220 08:15:36.625680 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"bccf823b97fcab3fd9a064d741fe0510b30d622a5a022d73e528e79d103d68bc"} Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.025207 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.025306 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.025377 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.026288 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a2c247070dfb4d0488751bb46e5543e834b3853c983d98f00e419c6aabfdeb7c"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.026394 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://a2c247070dfb4d0488751bb46e5543e834b3853c983d98f00e419c6aabfdeb7c" gracePeriod=600 Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.644699 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="a2c247070dfb4d0488751bb46e5543e834b3853c983d98f00e419c6aabfdeb7c" exitCode=0 Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.644752 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"a2c247070dfb4d0488751bb46e5543e834b3853c983d98f00e419c6aabfdeb7c"} Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.645463 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"b5405a03f834a339ac758e6efb1e31a8c094b3f6b973ea8540eefc716ba7bf7f"} Feb 20 08:15:38 crc kubenswrapper[4948]: I0220 08:15:38.645501 4948 scope.go:117] "RemoveContainer" containerID="6791802a782456e3c45df951fa55097ce0ea3affd6a9878fc2981511046d1b09" Feb 20 08:15:39 crc kubenswrapper[4948]: I0220 08:15:39.659604 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"15c07d823d4ea0c40991b497c78c8d77a94305232ecc7462264b9b70da7d7f0b"} Feb 20 08:15:41 crc kubenswrapper[4948]: I0220 08:15:41.678904 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" event={"ID":"0bfc877e-59ed-4803-b07a-b9787b1ab73d","Type":"ContainerStarted","Data":"e1ed772a07c900cbc9a9b6fa2cca2ea52073eff0653e8ed0b21f025559c1087e"} Feb 20 08:15:41 crc kubenswrapper[4948]: I0220 08:15:41.679517 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:41 crc kubenswrapper[4948]: I0220 08:15:41.720220 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" podStartSLOduration=7.720203128 podStartE2EDuration="7.720203128s" podCreationTimestamp="2026-02-20 08:15:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:15:41.714045724 +0000 UTC m=+590.688540574" watchObservedRunningTime="2026-02-20 08:15:41.720203128 +0000 UTC m=+590.694697948" Feb 20 08:15:41 crc kubenswrapper[4948]: I0220 08:15:41.747309 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:42 crc kubenswrapper[4948]: I0220 08:15:42.687887 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:42 crc kubenswrapper[4948]: I0220 08:15:42.688319 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:42 crc kubenswrapper[4948]: I0220 08:15:42.775082 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:15:47 crc kubenswrapper[4948]: I0220 08:15:47.723182 4948 scope.go:117] "RemoveContainer" containerID="4ecf8c7c9ed34410ff4619b18254e2eb90170911300e6cd6a7284e84e8937246" Feb 20 08:15:47 crc kubenswrapper[4948]: E0220 08:15:47.723669 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-x7s56_openshift-multus(b96124e4-0a74-4578-9142-fd728eb9f99e)\"" pod="openshift-multus/multus-x7s56" podUID="b96124e4-0a74-4578-9142-fd728eb9f99e" Feb 20 08:15:52 crc kubenswrapper[4948]: I0220 08:15:52.003509 4948 scope.go:117] "RemoveContainer" containerID="2f4093ec9d71181c58ad80a8c4ee3692148590df14aadaec96c0f54a843d1f8c" Feb 20 08:15:52 crc kubenswrapper[4948]: I0220 08:15:52.761470 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/2.log" Feb 20 08:15:58 crc kubenswrapper[4948]: I0220 08:15:58.723079 4948 scope.go:117] "RemoveContainer" containerID="4ecf8c7c9ed34410ff4619b18254e2eb90170911300e6cd6a7284e84e8937246" Feb 20 08:15:59 crc kubenswrapper[4948]: I0220 08:15:59.814731 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7s56_b96124e4-0a74-4578-9142-fd728eb9f99e/kube-multus/2.log" Feb 20 08:15:59 crc kubenswrapper[4948]: I0220 08:15:59.815102 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7s56" event={"ID":"b96124e4-0a74-4578-9142-fd728eb9f99e","Type":"ContainerStarted","Data":"e509f78ee7018faa8829b9687c5ed7446ec2bd36a0ef7fc02e9a2a538c8be56d"} Feb 20 08:16:05 crc kubenswrapper[4948]: I0220 08:16:05.307159 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cvhbs" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.077284 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w"] Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.078925 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.082225 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.087995 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w"] Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.204284 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvwb7\" (UniqueName: \"kubernetes.io/projected/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-kube-api-access-nvwb7\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.204381 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.204412 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.305491 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.305564 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.305684 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvwb7\" (UniqueName: \"kubernetes.io/projected/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-kube-api-access-nvwb7\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.306550 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.306741 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.338270 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvwb7\" (UniqueName: \"kubernetes.io/projected/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-kube-api-access-nvwb7\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.398273 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.673937 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w"] Feb 20 08:16:12 crc kubenswrapper[4948]: W0220 08:16:12.681901 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb3ec19c_541d_4b32_9f5c_ac61f28bdc2a.slice/crio-548722355d09a176b724709bdf45edc717f09882503ef4ddac12a3f79190d30d WatchSource:0}: Error finding container 548722355d09a176b724709bdf45edc717f09882503ef4ddac12a3f79190d30d: Status 404 returned error can't find the container with id 548722355d09a176b724709bdf45edc717f09882503ef4ddac12a3f79190d30d Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.898185 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" event={"ID":"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a","Type":"ContainerStarted","Data":"23679add441ce28e94c0d0fe925e7ecfaf8abe0638184164f4d2902ffb5cfa86"} Feb 20 08:16:12 crc kubenswrapper[4948]: I0220 08:16:12.898253 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" event={"ID":"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a","Type":"ContainerStarted","Data":"548722355d09a176b724709bdf45edc717f09882503ef4ddac12a3f79190d30d"} Feb 20 08:16:13 crc kubenswrapper[4948]: I0220 08:16:13.905260 4948 generic.go:334] "Generic (PLEG): container finished" podID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerID="23679add441ce28e94c0d0fe925e7ecfaf8abe0638184164f4d2902ffb5cfa86" exitCode=0 Feb 20 08:16:13 crc kubenswrapper[4948]: I0220 08:16:13.905317 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" event={"ID":"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a","Type":"ContainerDied","Data":"23679add441ce28e94c0d0fe925e7ecfaf8abe0638184164f4d2902ffb5cfa86"} Feb 20 08:16:15 crc kubenswrapper[4948]: I0220 08:16:15.922364 4948 generic.go:334] "Generic (PLEG): container finished" podID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerID="4b79219701d24d57bef90a1d915b063c7b1bfa1fad99516543195699165a1cc8" exitCode=0 Feb 20 08:16:15 crc kubenswrapper[4948]: I0220 08:16:15.922424 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" event={"ID":"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a","Type":"ContainerDied","Data":"4b79219701d24d57bef90a1d915b063c7b1bfa1fad99516543195699165a1cc8"} Feb 20 08:16:16 crc kubenswrapper[4948]: I0220 08:16:16.932125 4948 generic.go:334] "Generic (PLEG): container finished" podID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerID="8dd2cd73dd9334db76086a27f76ad986432450f60bbc24bf1cebe547d08c3e88" exitCode=0 Feb 20 08:16:16 crc kubenswrapper[4948]: I0220 08:16:16.932170 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" event={"ID":"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a","Type":"ContainerDied","Data":"8dd2cd73dd9334db76086a27f76ad986432450f60bbc24bf1cebe547d08c3e88"} Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.217368 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.391781 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-util\") pod \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.391880 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvwb7\" (UniqueName: \"kubernetes.io/projected/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-kube-api-access-nvwb7\") pod \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.391955 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-bundle\") pod \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\" (UID: \"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a\") " Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.393130 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-bundle" (OuterVolumeSpecName: "bundle") pod "bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" (UID: "bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.398721 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-kube-api-access-nvwb7" (OuterVolumeSpecName: "kube-api-access-nvwb7") pod "bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" (UID: "bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a"). InnerVolumeSpecName "kube-api-access-nvwb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.476391 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-util" (OuterVolumeSpecName: "util") pod "bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" (UID: "bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.493722 4948 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-util\") on node \"crc\" DevicePath \"\"" Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.493759 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvwb7\" (UniqueName: \"kubernetes.io/projected/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-kube-api-access-nvwb7\") on node \"crc\" DevicePath \"\"" Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.493780 4948 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.948617 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" event={"ID":"bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a","Type":"ContainerDied","Data":"548722355d09a176b724709bdf45edc717f09882503ef4ddac12a3f79190d30d"} Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.948675 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="548722355d09a176b724709bdf45edc717f09882503ef4ddac12a3f79190d30d" Feb 20 08:16:18 crc kubenswrapper[4948]: I0220 08:16:18.948783 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.023253 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-xqtvg"] Feb 20 08:16:21 crc kubenswrapper[4948]: E0220 08:16:21.023896 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerName="util" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.023910 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerName="util" Feb 20 08:16:21 crc kubenswrapper[4948]: E0220 08:16:21.023925 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerName="pull" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.023933 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerName="pull" Feb 20 08:16:21 crc kubenswrapper[4948]: E0220 08:16:21.023943 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerName="extract" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.023952 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerName="extract" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.024112 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a" containerName="extract" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.024535 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-xqtvg" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.026079 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w25ws\" (UniqueName: \"kubernetes.io/projected/59277281-0112-45da-a64a-1b1d02b90473-kube-api-access-w25ws\") pod \"nmstate-operator-694c9596b7-xqtvg\" (UID: \"59277281-0112-45da-a64a-1b1d02b90473\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-xqtvg" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.027247 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-vkzsw" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.027673 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.028449 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.036738 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-xqtvg"] Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.127445 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w25ws\" (UniqueName: \"kubernetes.io/projected/59277281-0112-45da-a64a-1b1d02b90473-kube-api-access-w25ws\") pod \"nmstate-operator-694c9596b7-xqtvg\" (UID: \"59277281-0112-45da-a64a-1b1d02b90473\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-xqtvg" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.151742 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w25ws\" (UniqueName: \"kubernetes.io/projected/59277281-0112-45da-a64a-1b1d02b90473-kube-api-access-w25ws\") pod \"nmstate-operator-694c9596b7-xqtvg\" (UID: \"59277281-0112-45da-a64a-1b1d02b90473\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-xqtvg" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.350483 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-xqtvg" Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.529077 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-xqtvg"] Feb 20 08:16:21 crc kubenswrapper[4948]: I0220 08:16:21.966606 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-xqtvg" event={"ID":"59277281-0112-45da-a64a-1b1d02b90473","Type":"ContainerStarted","Data":"c05b6a8309c306ed7cfbbb7e06e1c7f43f95c3a41836739a15a14819c2700731"} Feb 20 08:16:28 crc kubenswrapper[4948]: I0220 08:16:28.014849 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-xqtvg" event={"ID":"59277281-0112-45da-a64a-1b1d02b90473","Type":"ContainerStarted","Data":"4467fa79cf451a4a49018633dda15922cfc832112e00e679a559ad7e775fb518"} Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.059138 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-694c9596b7-xqtvg" podStartSLOduration=3.163767336 podStartE2EDuration="9.059112218s" podCreationTimestamp="2026-02-20 08:16:20 +0000 UTC" firstStartedPulling="2026-02-20 08:16:21.534283725 +0000 UTC m=+630.508778545" lastFinishedPulling="2026-02-20 08:16:27.429628597 +0000 UTC m=+636.404123427" observedRunningTime="2026-02-20 08:16:28.046165941 +0000 UTC m=+637.020660821" watchObservedRunningTime="2026-02-20 08:16:29.059112218 +0000 UTC m=+638.033607068" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.063696 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-tdst7"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.065120 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.067263 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-xqmp6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.097052 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-dxdh6"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.098626 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.102211 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-tdst7"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.113619 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.114775 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.126811 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.160775 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.212243 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.212902 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.215919 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.216207 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.222108 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-llbnp" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.224078 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.248727 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-dbus-socket\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.248791 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-ovs-socket\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.248841 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/da0e4284-f20d-44a8-b529-e388f15822d3-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-77sdv\" (UID: \"da0e4284-f20d-44a8-b529-e388f15822d3\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.248873 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67xdv\" (UniqueName: \"kubernetes.io/projected/3a869b3a-b757-49ba-9096-2e562d980aae-kube-api-access-67xdv\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.248900 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-nmstate-lock\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.248922 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp789\" (UniqueName: \"kubernetes.io/projected/da0e4284-f20d-44a8-b529-e388f15822d3-kube-api-access-pp789\") pod \"nmstate-webhook-866bcb46dc-77sdv\" (UID: \"da0e4284-f20d-44a8-b529-e388f15822d3\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.248943 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tgvg\" (UniqueName: \"kubernetes.io/projected/a7b34c06-d0c5-4f20-9521-eec500ea2510-kube-api-access-6tgvg\") pod \"nmstate-metrics-58c85c668d-tdst7\" (UID: \"a7b34c06-d0c5-4f20-9521-eec500ea2510\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.350559 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67xdv\" (UniqueName: \"kubernetes.io/projected/3a869b3a-b757-49ba-9096-2e562d980aae-kube-api-access-67xdv\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.350629 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqt4s\" (UniqueName: \"kubernetes.io/projected/47461ed8-c51b-4014-bcf0-4d95a0278f85-kube-api-access-jqt4s\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.350661 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-nmstate-lock\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.350695 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp789\" (UniqueName: \"kubernetes.io/projected/da0e4284-f20d-44a8-b529-e388f15822d3-kube-api-access-pp789\") pod \"nmstate-webhook-866bcb46dc-77sdv\" (UID: \"da0e4284-f20d-44a8-b529-e388f15822d3\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.350726 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tgvg\" (UniqueName: \"kubernetes.io/projected/a7b34c06-d0c5-4f20-9521-eec500ea2510-kube-api-access-6tgvg\") pod \"nmstate-metrics-58c85c668d-tdst7\" (UID: \"a7b34c06-d0c5-4f20-9521-eec500ea2510\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.350778 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-nmstate-lock\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.350790 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/47461ed8-c51b-4014-bcf0-4d95a0278f85-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.350945 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-dbus-socket\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.351023 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-ovs-socket\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.351066 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/47461ed8-c51b-4014-bcf0-4d95a0278f85-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.351097 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/da0e4284-f20d-44a8-b529-e388f15822d3-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-77sdv\" (UID: \"da0e4284-f20d-44a8-b529-e388f15822d3\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.351126 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-ovs-socket\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.351241 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3a869b3a-b757-49ba-9096-2e562d980aae-dbus-socket\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.361135 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/da0e4284-f20d-44a8-b529-e388f15822d3-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-77sdv\" (UID: \"da0e4284-f20d-44a8-b529-e388f15822d3\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.371190 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67xdv\" (UniqueName: \"kubernetes.io/projected/3a869b3a-b757-49ba-9096-2e562d980aae-kube-api-access-67xdv\") pod \"nmstate-handler-dxdh6\" (UID: \"3a869b3a-b757-49ba-9096-2e562d980aae\") " pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.375008 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tgvg\" (UniqueName: \"kubernetes.io/projected/a7b34c06-d0c5-4f20-9521-eec500ea2510-kube-api-access-6tgvg\") pod \"nmstate-metrics-58c85c668d-tdst7\" (UID: \"a7b34c06-d0c5-4f20-9521-eec500ea2510\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.381563 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp789\" (UniqueName: \"kubernetes.io/projected/da0e4284-f20d-44a8-b529-e388f15822d3-kube-api-access-pp789\") pod \"nmstate-webhook-866bcb46dc-77sdv\" (UID: \"da0e4284-f20d-44a8-b529-e388f15822d3\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.405336 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.428718 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.446026 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.449323 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-64dcb9c7f-fwdtz"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.450390 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.452118 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/47461ed8-c51b-4014-bcf0-4d95a0278f85-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.452187 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/47461ed8-c51b-4014-bcf0-4d95a0278f85-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.452234 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqt4s\" (UniqueName: \"kubernetes.io/projected/47461ed8-c51b-4014-bcf0-4d95a0278f85-kube-api-access-jqt4s\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.453470 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/47461ed8-c51b-4014-bcf0-4d95a0278f85-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.481196 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqt4s\" (UniqueName: \"kubernetes.io/projected/47461ed8-c51b-4014-bcf0-4d95a0278f85-kube-api-access-jqt4s\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.481219 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/47461ed8-c51b-4014-bcf0-4d95a0278f85-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-9mv9z\" (UID: \"47461ed8-c51b-4014-bcf0-4d95a0278f85\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.486852 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-64dcb9c7f-fwdtz"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.533756 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.553420 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-config\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.553477 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhtt6\" (UniqueName: \"kubernetes.io/projected/9b67f523-c74e-4b03-99cf-64fe0dd981f2-kube-api-access-mhtt6\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.553575 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-trusted-ca-bundle\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.553648 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-oauth-serving-cert\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.553684 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-service-ca\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.553933 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-serving-cert\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.554036 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-oauth-config\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.639508 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-tdst7"] Feb 20 08:16:29 crc kubenswrapper[4948]: W0220 08:16:29.653229 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7b34c06_d0c5_4f20_9521_eec500ea2510.slice/crio-2f69a3cdfed473617a7a51a1fd6975716f998a88d66b252db4a98122759defcd WatchSource:0}: Error finding container 2f69a3cdfed473617a7a51a1fd6975716f998a88d66b252db4a98122759defcd: Status 404 returned error can't find the container with id 2f69a3cdfed473617a7a51a1fd6975716f998a88d66b252db4a98122759defcd Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.654805 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhtt6\" (UniqueName: \"kubernetes.io/projected/9b67f523-c74e-4b03-99cf-64fe0dd981f2-kube-api-access-mhtt6\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.654848 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-trusted-ca-bundle\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.654876 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-oauth-serving-cert\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.654895 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-service-ca\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.654939 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-serving-cert\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.654982 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-oauth-config\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.655002 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-config\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.656575 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-oauth-serving-cert\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.656966 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-trusted-ca-bundle\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.657070 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-config\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.657733 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9b67f523-c74e-4b03-99cf-64fe0dd981f2-service-ca\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.659873 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-serving-cert\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.669586 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9b67f523-c74e-4b03-99cf-64fe0dd981f2-console-oauth-config\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.673944 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhtt6\" (UniqueName: \"kubernetes.io/projected/9b67f523-c74e-4b03-99cf-64fe0dd981f2-kube-api-access-mhtt6\") pod \"console-64dcb9c7f-fwdtz\" (UID: \"9b67f523-c74e-4b03-99cf-64fe0dd981f2\") " pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: W0220 08:16:29.781794 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47461ed8_c51b_4014_bcf0_4d95a0278f85.slice/crio-7d31361835d89022a27dbe1075d24c03db175adbb0decc2479704bd81951c240 WatchSource:0}: Error finding container 7d31361835d89022a27dbe1075d24c03db175adbb0decc2479704bd81951c240: Status 404 returned error can't find the container with id 7d31361835d89022a27dbe1075d24c03db175adbb0decc2479704bd81951c240 Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.782771 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z"] Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.804516 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:29 crc kubenswrapper[4948]: I0220 08:16:29.950793 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv"] Feb 20 08:16:29 crc kubenswrapper[4948]: W0220 08:16:29.959403 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda0e4284_f20d_44a8_b529_e388f15822d3.slice/crio-b8fa1601d3f454168f034271bfc5dc2d68a4cc6611e8568c6cff0f11fb9e443d WatchSource:0}: Error finding container b8fa1601d3f454168f034271bfc5dc2d68a4cc6611e8568c6cff0f11fb9e443d: Status 404 returned error can't find the container with id b8fa1601d3f454168f034271bfc5dc2d68a4cc6611e8568c6cff0f11fb9e443d Feb 20 08:16:30 crc kubenswrapper[4948]: I0220 08:16:30.029106 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" event={"ID":"47461ed8-c51b-4014-bcf0-4d95a0278f85","Type":"ContainerStarted","Data":"7d31361835d89022a27dbe1075d24c03db175adbb0decc2479704bd81951c240"} Feb 20 08:16:30 crc kubenswrapper[4948]: I0220 08:16:30.030708 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" event={"ID":"a7b34c06-d0c5-4f20-9521-eec500ea2510","Type":"ContainerStarted","Data":"2f69a3cdfed473617a7a51a1fd6975716f998a88d66b252db4a98122759defcd"} Feb 20 08:16:30 crc kubenswrapper[4948]: I0220 08:16:30.031788 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" event={"ID":"da0e4284-f20d-44a8-b529-e388f15822d3","Type":"ContainerStarted","Data":"b8fa1601d3f454168f034271bfc5dc2d68a4cc6611e8568c6cff0f11fb9e443d"} Feb 20 08:16:30 crc kubenswrapper[4948]: I0220 08:16:30.033716 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-dxdh6" event={"ID":"3a869b3a-b757-49ba-9096-2e562d980aae","Type":"ContainerStarted","Data":"3c9a321ad1b783cb64857b52dd5a94a4cc043894390bb304392d324d8553e8fc"} Feb 20 08:16:30 crc kubenswrapper[4948]: I0220 08:16:30.058283 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-64dcb9c7f-fwdtz"] Feb 20 08:16:30 crc kubenswrapper[4948]: W0220 08:16:30.063130 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b67f523_c74e_4b03_99cf_64fe0dd981f2.slice/crio-b7245036c705e569134bea807738097ccff83132fb89b9e2861c36fcfe1ef775 WatchSource:0}: Error finding container b7245036c705e569134bea807738097ccff83132fb89b9e2861c36fcfe1ef775: Status 404 returned error can't find the container with id b7245036c705e569134bea807738097ccff83132fb89b9e2861c36fcfe1ef775 Feb 20 08:16:31 crc kubenswrapper[4948]: I0220 08:16:31.042139 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-64dcb9c7f-fwdtz" event={"ID":"9b67f523-c74e-4b03-99cf-64fe0dd981f2","Type":"ContainerStarted","Data":"5b4b88a4af63b9a4d22a23d19cc8005963d21d10edd08aa88ee963fbb6f2a798"} Feb 20 08:16:31 crc kubenswrapper[4948]: I0220 08:16:31.042440 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-64dcb9c7f-fwdtz" event={"ID":"9b67f523-c74e-4b03-99cf-64fe0dd981f2","Type":"ContainerStarted","Data":"b7245036c705e569134bea807738097ccff83132fb89b9e2861c36fcfe1ef775"} Feb 20 08:16:31 crc kubenswrapper[4948]: I0220 08:16:31.756591 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-64dcb9c7f-fwdtz" podStartSLOduration=2.756564242 podStartE2EDuration="2.756564242s" podCreationTimestamp="2026-02-20 08:16:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:16:31.06452074 +0000 UTC m=+640.039015560" watchObservedRunningTime="2026-02-20 08:16:31.756564242 +0000 UTC m=+640.731059102" Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.059049 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" event={"ID":"47461ed8-c51b-4014-bcf0-4d95a0278f85","Type":"ContainerStarted","Data":"d126b06cf7663226603898b9659ec53aa29323f2f8277118fa53815a5cd83895"} Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.070930 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" event={"ID":"a7b34c06-d0c5-4f20-9521-eec500ea2510","Type":"ContainerStarted","Data":"690f316803c2ecde318a6e40f63ad3ba7cd1be8c5b0541c5f60be03c437f23d3"} Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.073499 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" event={"ID":"da0e4284-f20d-44a8-b529-e388f15822d3","Type":"ContainerStarted","Data":"30d1210c199eb6e76765b57e55c646439da7610a9659e99f3f03629f2824d553"} Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.073896 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.077387 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-dxdh6" event={"ID":"3a869b3a-b757-49ba-9096-2e562d980aae","Type":"ContainerStarted","Data":"6e3aa8eb0ab95c174b74ca9e6b49ff3ed3735f50ca9cd7ba5ed209d3e8bfcf59"} Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.078046 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.085361 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-9mv9z" podStartSLOduration=1.5106040200000002 podStartE2EDuration="4.085344586s" podCreationTimestamp="2026-02-20 08:16:29 +0000 UTC" firstStartedPulling="2026-02-20 08:16:29.784347201 +0000 UTC m=+638.758842011" lastFinishedPulling="2026-02-20 08:16:32.359087757 +0000 UTC m=+641.333582577" observedRunningTime="2026-02-20 08:16:33.084569997 +0000 UTC m=+642.059064827" watchObservedRunningTime="2026-02-20 08:16:33.085344586 +0000 UTC m=+642.059839446" Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.116356 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" podStartSLOduration=1.729020261 podStartE2EDuration="4.116331261s" podCreationTimestamp="2026-02-20 08:16:29 +0000 UTC" firstStartedPulling="2026-02-20 08:16:29.962638169 +0000 UTC m=+638.937132999" lastFinishedPulling="2026-02-20 08:16:32.349949139 +0000 UTC m=+641.324443999" observedRunningTime="2026-02-20 08:16:33.111405518 +0000 UTC m=+642.085900368" watchObservedRunningTime="2026-02-20 08:16:33.116331261 +0000 UTC m=+642.090826091" Feb 20 08:16:33 crc kubenswrapper[4948]: I0220 08:16:33.135387 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-dxdh6" podStartSLOduration=1.277670536 podStartE2EDuration="4.135368527s" podCreationTimestamp="2026-02-20 08:16:29 +0000 UTC" firstStartedPulling="2026-02-20 08:16:29.504666378 +0000 UTC m=+638.479161198" lastFinishedPulling="2026-02-20 08:16:32.362364369 +0000 UTC m=+641.336859189" observedRunningTime="2026-02-20 08:16:33.128027203 +0000 UTC m=+642.102522063" watchObservedRunningTime="2026-02-20 08:16:33.135368527 +0000 UTC m=+642.109863367" Feb 20 08:16:35 crc kubenswrapper[4948]: I0220 08:16:35.094569 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" event={"ID":"a7b34c06-d0c5-4f20-9521-eec500ea2510","Type":"ContainerStarted","Data":"fd7928ff068e20efed3dba24766aca8d8b2797500a32b05fdb406cc49dcc9cab"} Feb 20 08:16:35 crc kubenswrapper[4948]: I0220 08:16:35.119804 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58c85c668d-tdst7" podStartSLOduration=1.2550912300000001 podStartE2EDuration="6.119788593s" podCreationTimestamp="2026-02-20 08:16:29 +0000 UTC" firstStartedPulling="2026-02-20 08:16:29.657638533 +0000 UTC m=+638.632133353" lastFinishedPulling="2026-02-20 08:16:34.522335896 +0000 UTC m=+643.496830716" observedRunningTime="2026-02-20 08:16:35.11568264 +0000 UTC m=+644.090177460" watchObservedRunningTime="2026-02-20 08:16:35.119788593 +0000 UTC m=+644.094283413" Feb 20 08:16:39 crc kubenswrapper[4948]: I0220 08:16:39.468639 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-dxdh6" Feb 20 08:16:39 crc kubenswrapper[4948]: I0220 08:16:39.804644 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:39 crc kubenswrapper[4948]: I0220 08:16:39.804725 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:39 crc kubenswrapper[4948]: I0220 08:16:39.810604 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:40 crc kubenswrapper[4948]: I0220 08:16:40.134277 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-64dcb9c7f-fwdtz" Feb 20 08:16:40 crc kubenswrapper[4948]: I0220 08:16:40.194462 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-k8zcr"] Feb 20 08:16:49 crc kubenswrapper[4948]: I0220 08:16:49.457172 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-77sdv" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.263137 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-k8zcr" podUID="f81366e5-fbdb-4a0b-a91f-f603d9b8570d" containerName="console" containerID="cri-o://3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78" gracePeriod=15 Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.597144 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-k8zcr_f81366e5-fbdb-4a0b-a91f-f603d9b8570d/console/0.log" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.597493 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.627421 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct"] Feb 20 08:17:05 crc kubenswrapper[4948]: E0220 08:17:05.627656 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f81366e5-fbdb-4a0b-a91f-f603d9b8570d" containerName="console" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.627667 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f81366e5-fbdb-4a0b-a91f-f603d9b8570d" containerName="console" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.627768 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f81366e5-fbdb-4a0b-a91f-f603d9b8570d" containerName="console" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.628445 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.630408 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.637416 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct"] Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.789351 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-oauth-serving-cert\") pod \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.789500 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7b5rg\" (UniqueName: \"kubernetes.io/projected/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-kube-api-access-7b5rg\") pod \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.789542 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-trusted-ca-bundle\") pod \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.789580 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-config\") pod \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.789669 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-service-ca\") pod \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.789718 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-serving-cert\") pod \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.789796 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-oauth-config\") pod \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\" (UID: \"f81366e5-fbdb-4a0b-a91f-f603d9b8570d\") " Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.790063 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.790132 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcj8g\" (UniqueName: \"kubernetes.io/projected/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-kube-api-access-lcj8g\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.790217 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.790826 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-config" (OuterVolumeSpecName: "console-config") pod "f81366e5-fbdb-4a0b-a91f-f603d9b8570d" (UID: "f81366e5-fbdb-4a0b-a91f-f603d9b8570d"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.791014 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-service-ca" (OuterVolumeSpecName: "service-ca") pod "f81366e5-fbdb-4a0b-a91f-f603d9b8570d" (UID: "f81366e5-fbdb-4a0b-a91f-f603d9b8570d"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.791637 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "f81366e5-fbdb-4a0b-a91f-f603d9b8570d" (UID: "f81366e5-fbdb-4a0b-a91f-f603d9b8570d"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.791690 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "f81366e5-fbdb-4a0b-a91f-f603d9b8570d" (UID: "f81366e5-fbdb-4a0b-a91f-f603d9b8570d"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.801495 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "f81366e5-fbdb-4a0b-a91f-f603d9b8570d" (UID: "f81366e5-fbdb-4a0b-a91f-f603d9b8570d"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.801682 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-kube-api-access-7b5rg" (OuterVolumeSpecName: "kube-api-access-7b5rg") pod "f81366e5-fbdb-4a0b-a91f-f603d9b8570d" (UID: "f81366e5-fbdb-4a0b-a91f-f603d9b8570d"). InnerVolumeSpecName "kube-api-access-7b5rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.802277 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "f81366e5-fbdb-4a0b-a91f-f603d9b8570d" (UID: "f81366e5-fbdb-4a0b-a91f-f603d9b8570d"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.891558 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.891656 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcj8g\" (UniqueName: \"kubernetes.io/projected/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-kube-api-access-lcj8g\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892006 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892150 4948 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892247 4948 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892274 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7b5rg\" (UniqueName: \"kubernetes.io/projected/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-kube-api-access-7b5rg\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892342 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892375 4948 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892465 4948 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892546 4948 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f81366e5-fbdb-4a0b-a91f-f603d9b8570d-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892394 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.892410 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.914180 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcj8g\" (UniqueName: \"kubernetes.io/projected/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-kube-api-access-lcj8g\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:05 crc kubenswrapper[4948]: I0220 08:17:05.945367 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.147253 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct"] Feb 20 08:17:06 crc kubenswrapper[4948]: W0220 08:17:06.156401 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e5e2f81_8ea9_4ddb_b2ab_584c40b15916.slice/crio-b58d77869abc6a06a40c2616d2fffd5dda5671eb4c1eccfaea6ec70a3e40332b WatchSource:0}: Error finding container b58d77869abc6a06a40c2616d2fffd5dda5671eb4c1eccfaea6ec70a3e40332b: Status 404 returned error can't find the container with id b58d77869abc6a06a40c2616d2fffd5dda5671eb4c1eccfaea6ec70a3e40332b Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.349266 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-k8zcr_f81366e5-fbdb-4a0b-a91f-f603d9b8570d/console/0.log" Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.349351 4948 generic.go:334] "Generic (PLEG): container finished" podID="f81366e5-fbdb-4a0b-a91f-f603d9b8570d" containerID="3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78" exitCode=2 Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.349449 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k8zcr" event={"ID":"f81366e5-fbdb-4a0b-a91f-f603d9b8570d","Type":"ContainerDied","Data":"3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78"} Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.349493 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k8zcr" event={"ID":"f81366e5-fbdb-4a0b-a91f-f603d9b8570d","Type":"ContainerDied","Data":"61712307e24396811f0211771ba686b9bbc2eb0844719329e63b16d44d321de6"} Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.349530 4948 scope.go:117] "RemoveContainer" containerID="3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78" Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.349570 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k8zcr" Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.353281 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" event={"ID":"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916","Type":"ContainerStarted","Data":"5be15e1279406c468ca430c5a05ee089f20fb99bdd79d38237b6ed24ccf2f18c"} Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.353350 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" event={"ID":"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916","Type":"ContainerStarted","Data":"b58d77869abc6a06a40c2616d2fffd5dda5671eb4c1eccfaea6ec70a3e40332b"} Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.390548 4948 scope.go:117] "RemoveContainer" containerID="3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78" Feb 20 08:17:06 crc kubenswrapper[4948]: E0220 08:17:06.391517 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78\": container with ID starting with 3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78 not found: ID does not exist" containerID="3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78" Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.391595 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78"} err="failed to get container status \"3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78\": rpc error: code = NotFound desc = could not find container \"3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78\": container with ID starting with 3745ecb9e997c9c1a842fb71042af1df4c9828447494e128f09e0903f609bd78 not found: ID does not exist" Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.433209 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-k8zcr"] Feb 20 08:17:06 crc kubenswrapper[4948]: I0220 08:17:06.435588 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-k8zcr"] Feb 20 08:17:07 crc kubenswrapper[4948]: I0220 08:17:07.367371 4948 generic.go:334] "Generic (PLEG): container finished" podID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerID="5be15e1279406c468ca430c5a05ee089f20fb99bdd79d38237b6ed24ccf2f18c" exitCode=0 Feb 20 08:17:07 crc kubenswrapper[4948]: I0220 08:17:07.367471 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" event={"ID":"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916","Type":"ContainerDied","Data":"5be15e1279406c468ca430c5a05ee089f20fb99bdd79d38237b6ed24ccf2f18c"} Feb 20 08:17:07 crc kubenswrapper[4948]: I0220 08:17:07.736770 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f81366e5-fbdb-4a0b-a91f-f603d9b8570d" path="/var/lib/kubelet/pods/f81366e5-fbdb-4a0b-a91f-f603d9b8570d/volumes" Feb 20 08:17:09 crc kubenswrapper[4948]: I0220 08:17:09.386143 4948 generic.go:334] "Generic (PLEG): container finished" podID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerID="dfd6183de2e06d2690aa7e1da7e9625df33f176428e6c2492950a7e9940ee6df" exitCode=0 Feb 20 08:17:09 crc kubenswrapper[4948]: I0220 08:17:09.386216 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" event={"ID":"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916","Type":"ContainerDied","Data":"dfd6183de2e06d2690aa7e1da7e9625df33f176428e6c2492950a7e9940ee6df"} Feb 20 08:17:10 crc kubenswrapper[4948]: I0220 08:17:10.397312 4948 generic.go:334] "Generic (PLEG): container finished" podID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerID="cfec89ba0a0c86027899432b1991c477117822e794036be26b25129aec08acc2" exitCode=0 Feb 20 08:17:10 crc kubenswrapper[4948]: I0220 08:17:10.397378 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" event={"ID":"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916","Type":"ContainerDied","Data":"cfec89ba0a0c86027899432b1991c477117822e794036be26b25129aec08acc2"} Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.753541 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.879883 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcj8g\" (UniqueName: \"kubernetes.io/projected/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-kube-api-access-lcj8g\") pod \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.880065 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-bundle\") pod \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.880193 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-util\") pod \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\" (UID: \"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916\") " Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.882197 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-bundle" (OuterVolumeSpecName: "bundle") pod "9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" (UID: "9e5e2f81-8ea9-4ddb-b2ab-584c40b15916"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.890144 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-kube-api-access-lcj8g" (OuterVolumeSpecName: "kube-api-access-lcj8g") pod "9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" (UID: "9e5e2f81-8ea9-4ddb-b2ab-584c40b15916"). InnerVolumeSpecName "kube-api-access-lcj8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.905220 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-util" (OuterVolumeSpecName: "util") pod "9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" (UID: "9e5e2f81-8ea9-4ddb-b2ab-584c40b15916"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.981815 4948 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.981907 4948 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-util\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:11 crc kubenswrapper[4948]: I0220 08:17:11.981962 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcj8g\" (UniqueName: \"kubernetes.io/projected/9e5e2f81-8ea9-4ddb-b2ab-584c40b15916-kube-api-access-lcj8g\") on node \"crc\" DevicePath \"\"" Feb 20 08:17:12 crc kubenswrapper[4948]: I0220 08:17:12.415143 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" event={"ID":"9e5e2f81-8ea9-4ddb-b2ab-584c40b15916","Type":"ContainerDied","Data":"b58d77869abc6a06a40c2616d2fffd5dda5671eb4c1eccfaea6ec70a3e40332b"} Feb 20 08:17:12 crc kubenswrapper[4948]: I0220 08:17:12.415203 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b58d77869abc6a06a40c2616d2fffd5dda5671eb4c1eccfaea6ec70a3e40332b" Feb 20 08:17:12 crc kubenswrapper[4948]: I0220 08:17:12.415261 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.582734 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs"] Feb 20 08:17:20 crc kubenswrapper[4948]: E0220 08:17:20.583515 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerName="extract" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.583527 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerName="extract" Feb 20 08:17:20 crc kubenswrapper[4948]: E0220 08:17:20.583543 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerName="pull" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.583549 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerName="pull" Feb 20 08:17:20 crc kubenswrapper[4948]: E0220 08:17:20.583559 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerName="util" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.583565 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerName="util" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.583665 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5e2f81-8ea9-4ddb-b2ab-584c40b15916" containerName="extract" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.584055 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.587549 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.587654 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.587713 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.587851 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-pqnxf" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.587860 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.603193 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs"] Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.696127 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dncwx\" (UniqueName: \"kubernetes.io/projected/b4f86a2e-15f9-441c-953b-49f331d4122e-kube-api-access-dncwx\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.696182 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b4f86a2e-15f9-441c-953b-49f331d4122e-apiservice-cert\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.696273 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b4f86a2e-15f9-441c-953b-49f331d4122e-webhook-cert\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.797410 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dncwx\" (UniqueName: \"kubernetes.io/projected/b4f86a2e-15f9-441c-953b-49f331d4122e-kube-api-access-dncwx\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.797468 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b4f86a2e-15f9-441c-953b-49f331d4122e-apiservice-cert\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.797741 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b4f86a2e-15f9-441c-953b-49f331d4122e-webhook-cert\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.803749 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b4f86a2e-15f9-441c-953b-49f331d4122e-apiservice-cert\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.803808 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b4f86a2e-15f9-441c-953b-49f331d4122e-webhook-cert\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.825525 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dncwx\" (UniqueName: \"kubernetes.io/projected/b4f86a2e-15f9-441c-953b-49f331d4122e-kube-api-access-dncwx\") pod \"metallb-operator-controller-manager-8478c6b8cc-v98zs\" (UID: \"b4f86a2e-15f9-441c-953b-49f331d4122e\") " pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.899759 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.915074 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs"] Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.916114 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.919633 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.923139 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.930244 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-qc469" Feb 20 08:17:20 crc kubenswrapper[4948]: I0220 08:17:20.951916 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs"] Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.000398 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz9g7\" (UniqueName: \"kubernetes.io/projected/07022b13-7dc1-45d1-a626-91792bd3aa90-kube-api-access-dz9g7\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.000472 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/07022b13-7dc1-45d1-a626-91792bd3aa90-apiservice-cert\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.000602 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/07022b13-7dc1-45d1-a626-91792bd3aa90-webhook-cert\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.101851 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/07022b13-7dc1-45d1-a626-91792bd3aa90-webhook-cert\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.102309 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz9g7\" (UniqueName: \"kubernetes.io/projected/07022b13-7dc1-45d1-a626-91792bd3aa90-kube-api-access-dz9g7\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.102354 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/07022b13-7dc1-45d1-a626-91792bd3aa90-apiservice-cert\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.107074 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/07022b13-7dc1-45d1-a626-91792bd3aa90-apiservice-cert\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.121181 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/07022b13-7dc1-45d1-a626-91792bd3aa90-webhook-cert\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.121278 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz9g7\" (UniqueName: \"kubernetes.io/projected/07022b13-7dc1-45d1-a626-91792bd3aa90-kube-api-access-dz9g7\") pod \"metallb-operator-webhook-server-5fcd7d546c-kgpxs\" (UID: \"07022b13-7dc1-45d1-a626-91792bd3aa90\") " pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.147585 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs"] Feb 20 08:17:21 crc kubenswrapper[4948]: W0220 08:17:21.154745 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4f86a2e_15f9_441c_953b_49f331d4122e.slice/crio-fa52bf8c7effb4b51c609222f52518944307009826752b172359108a4af822c0 WatchSource:0}: Error finding container fa52bf8c7effb4b51c609222f52518944307009826752b172359108a4af822c0: Status 404 returned error can't find the container with id fa52bf8c7effb4b51c609222f52518944307009826752b172359108a4af822c0 Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.285261 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.468779 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" event={"ID":"b4f86a2e-15f9-441c-953b-49f331d4122e","Type":"ContainerStarted","Data":"fa52bf8c7effb4b51c609222f52518944307009826752b172359108a4af822c0"} Feb 20 08:17:21 crc kubenswrapper[4948]: I0220 08:17:21.499220 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs"] Feb 20 08:17:21 crc kubenswrapper[4948]: W0220 08:17:21.506015 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07022b13_7dc1_45d1_a626_91792bd3aa90.slice/crio-6500653b5f55a54ab5eb26076b35eb76477dc2e1b4770013d2121d1bf1f20b85 WatchSource:0}: Error finding container 6500653b5f55a54ab5eb26076b35eb76477dc2e1b4770013d2121d1bf1f20b85: Status 404 returned error can't find the container with id 6500653b5f55a54ab5eb26076b35eb76477dc2e1b4770013d2121d1bf1f20b85 Feb 20 08:17:22 crc kubenswrapper[4948]: I0220 08:17:22.474680 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" event={"ID":"07022b13-7dc1-45d1-a626-91792bd3aa90","Type":"ContainerStarted","Data":"6500653b5f55a54ab5eb26076b35eb76477dc2e1b4770013d2121d1bf1f20b85"} Feb 20 08:17:26 crc kubenswrapper[4948]: I0220 08:17:26.497914 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" event={"ID":"07022b13-7dc1-45d1-a626-91792bd3aa90","Type":"ContainerStarted","Data":"841867153dd83886e9618e9a4fae6c87b91836589203878bc1a8db173192ca25"} Feb 20 08:17:26 crc kubenswrapper[4948]: I0220 08:17:26.500303 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" event={"ID":"b4f86a2e-15f9-441c-953b-49f331d4122e","Type":"ContainerStarted","Data":"ab6d5674d550742387ada1d66fe00576345be4280789d0114664533ce4ca4a49"} Feb 20 08:17:26 crc kubenswrapper[4948]: I0220 08:17:26.500459 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:17:26 crc kubenswrapper[4948]: I0220 08:17:26.522012 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" podStartSLOduration=2.127778239 podStartE2EDuration="6.521985757s" podCreationTimestamp="2026-02-20 08:17:20 +0000 UTC" firstStartedPulling="2026-02-20 08:17:21.509822408 +0000 UTC m=+690.484317228" lastFinishedPulling="2026-02-20 08:17:25.904029926 +0000 UTC m=+694.878524746" observedRunningTime="2026-02-20 08:17:26.516051128 +0000 UTC m=+695.490545988" watchObservedRunningTime="2026-02-20 08:17:26.521985757 +0000 UTC m=+695.496480587" Feb 20 08:17:26 crc kubenswrapper[4948]: I0220 08:17:26.548203 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" podStartSLOduration=1.891160094 podStartE2EDuration="6.548183902s" podCreationTimestamp="2026-02-20 08:17:20 +0000 UTC" firstStartedPulling="2026-02-20 08:17:21.15667996 +0000 UTC m=+690.131174780" lastFinishedPulling="2026-02-20 08:17:25.813703768 +0000 UTC m=+694.788198588" observedRunningTime="2026-02-20 08:17:26.54532811 +0000 UTC m=+695.519822930" watchObservedRunningTime="2026-02-20 08:17:26.548183902 +0000 UTC m=+695.522678732" Feb 20 08:17:27 crc kubenswrapper[4948]: I0220 08:17:27.507158 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:17:38 crc kubenswrapper[4948]: I0220 08:17:38.024801 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:17:38 crc kubenswrapper[4948]: I0220 08:17:38.027058 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:17:41 crc kubenswrapper[4948]: I0220 08:17:41.296542 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5fcd7d546c-kgpxs" Feb 20 08:18:00 crc kubenswrapper[4948]: I0220 08:18:00.903364 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-8478c6b8cc-v98zs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.592153 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-k66nh"] Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.595525 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.600493 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.600507 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.601319 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-xnrzs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.616676 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs"] Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.616964 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics-certs\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.617058 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-reloader\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.617099 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-startup\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.617136 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-conf\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.617240 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.617297 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-sockets\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.617610 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q765x\" (UniqueName: \"kubernetes.io/projected/0131d05b-619d-43d8-aa38-ad6e9ce52e11-kube-api-access-q765x\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.617866 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.619878 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.635215 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs"] Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.680748 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-h7xhj"] Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.681641 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.683320 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.684462 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.684722 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.684900 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-fpnk6" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.713418 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-69bbfbf88f-c6sqk"] Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.714177 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.716412 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.718455 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.718497 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2c199acd-2fd3-4b9f-b50c-a266191c4777-metallb-excludel2\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.718527 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.718555 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-sockets\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.718607 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2506d4e2-b434-4fe4-970e-7cd14601677d-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-7zsxs\" (UID: \"2506d4e2-b434-4fe4-970e-7cd14601677d\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719138 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719159 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q765x\" (UniqueName: \"kubernetes.io/projected/0131d05b-619d-43d8-aa38-ad6e9ce52e11-kube-api-access-q765x\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719160 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-sockets\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719239 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics-certs\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719291 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-reloader\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719327 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-metrics-certs\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719361 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-startup\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719391 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-conf\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: E0220 08:18:01.719591 4948 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719659 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-reloader\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719781 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-conf\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: E0220 08:18:01.719892 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics-certs podName:0131d05b-619d-43d8-aa38-ad6e9ce52e11 nodeName:}" failed. No retries permitted until 2026-02-20 08:18:02.219872578 +0000 UTC m=+731.194367398 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics-certs") pod "frr-k8s-k66nh" (UID: "0131d05b-619d-43d8-aa38-ad6e9ce52e11") : secret "frr-k8s-certs-secret" not found Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.719922 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqst8\" (UniqueName: \"kubernetes.io/projected/2c199acd-2fd3-4b9f-b50c-a266191c4777-kube-api-access-dqst8\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.720005 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgh5n\" (UniqueName: \"kubernetes.io/projected/2506d4e2-b434-4fe4-970e-7cd14601677d-kube-api-access-wgh5n\") pod \"frr-k8s-webhook-server-78b44bf5bb-7zsxs\" (UID: \"2506d4e2-b434-4fe4-970e-7cd14601677d\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.720901 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0131d05b-619d-43d8-aa38-ad6e9ce52e11-frr-startup\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.732457 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-c6sqk"] Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.761208 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q765x\" (UniqueName: \"kubernetes.io/projected/0131d05b-619d-43d8-aa38-ad6e9ce52e11-kube-api-access-q765x\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.820890 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-metrics-certs\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.820948 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqst8\" (UniqueName: \"kubernetes.io/projected/2c199acd-2fd3-4b9f-b50c-a266191c4777-kube-api-access-dqst8\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.820975 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68ee0185-13d4-45c7-9a39-bb474e9314d0-cert\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.821011 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgh5n\" (UniqueName: \"kubernetes.io/projected/2506d4e2-b434-4fe4-970e-7cd14601677d-kube-api-access-wgh5n\") pod \"frr-k8s-webhook-server-78b44bf5bb-7zsxs\" (UID: \"2506d4e2-b434-4fe4-970e-7cd14601677d\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.821029 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.821046 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2c199acd-2fd3-4b9f-b50c-a266191c4777-metallb-excludel2\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.821066 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68ee0185-13d4-45c7-9a39-bb474e9314d0-metrics-certs\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: E0220 08:18:01.821066 4948 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.821100 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2506d4e2-b434-4fe4-970e-7cd14601677d-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-7zsxs\" (UID: \"2506d4e2-b434-4fe4-970e-7cd14601677d\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.821123 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxtfc\" (UniqueName: \"kubernetes.io/projected/68ee0185-13d4-45c7-9a39-bb474e9314d0-kube-api-access-qxtfc\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: E0220 08:18:01.821137 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-metrics-certs podName:2c199acd-2fd3-4b9f-b50c-a266191c4777 nodeName:}" failed. No retries permitted until 2026-02-20 08:18:02.32111968 +0000 UTC m=+731.295614500 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-metrics-certs") pod "speaker-h7xhj" (UID: "2c199acd-2fd3-4b9f-b50c-a266191c4777") : secret "speaker-certs-secret" not found Feb 20 08:18:01 crc kubenswrapper[4948]: E0220 08:18:01.821520 4948 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 20 08:18:01 crc kubenswrapper[4948]: E0220 08:18:01.821569 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist podName:2c199acd-2fd3-4b9f-b50c-a266191c4777 nodeName:}" failed. No retries permitted until 2026-02-20 08:18:02.32155485 +0000 UTC m=+731.296049670 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist") pod "speaker-h7xhj" (UID: "2c199acd-2fd3-4b9f-b50c-a266191c4777") : secret "metallb-memberlist" not found Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.821702 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2c199acd-2fd3-4b9f-b50c-a266191c4777-metallb-excludel2\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.826212 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2506d4e2-b434-4fe4-970e-7cd14601677d-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-7zsxs\" (UID: \"2506d4e2-b434-4fe4-970e-7cd14601677d\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.842506 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqst8\" (UniqueName: \"kubernetes.io/projected/2c199acd-2fd3-4b9f-b50c-a266191c4777-kube-api-access-dqst8\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.853742 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgh5n\" (UniqueName: \"kubernetes.io/projected/2506d4e2-b434-4fe4-970e-7cd14601677d-kube-api-access-wgh5n\") pod \"frr-k8s-webhook-server-78b44bf5bb-7zsxs\" (UID: \"2506d4e2-b434-4fe4-970e-7cd14601677d\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.922031 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68ee0185-13d4-45c7-9a39-bb474e9314d0-cert\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.922722 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68ee0185-13d4-45c7-9a39-bb474e9314d0-metrics-certs\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.922900 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxtfc\" (UniqueName: \"kubernetes.io/projected/68ee0185-13d4-45c7-9a39-bb474e9314d0-kube-api-access-qxtfc\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.925056 4948 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.926650 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68ee0185-13d4-45c7-9a39-bb474e9314d0-metrics-certs\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.935525 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68ee0185-13d4-45c7-9a39-bb474e9314d0-cert\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.938506 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:01 crc kubenswrapper[4948]: I0220 08:18:01.955015 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxtfc\" (UniqueName: \"kubernetes.io/projected/68ee0185-13d4-45c7-9a39-bb474e9314d0-kube-api-access-qxtfc\") pod \"controller-69bbfbf88f-c6sqk\" (UID: \"68ee0185-13d4-45c7-9a39-bb474e9314d0\") " pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.026507 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.227070 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics-certs\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.235070 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0131d05b-619d-43d8-aa38-ad6e9ce52e11-metrics-certs\") pod \"frr-k8s-k66nh\" (UID: \"0131d05b-619d-43d8-aa38-ad6e9ce52e11\") " pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.329550 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-metrics-certs\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.329736 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:02 crc kubenswrapper[4948]: E0220 08:18:02.329903 4948 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 20 08:18:02 crc kubenswrapper[4948]: E0220 08:18:02.330006 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist podName:2c199acd-2fd3-4b9f-b50c-a266191c4777 nodeName:}" failed. No retries permitted until 2026-02-20 08:18:03.329952892 +0000 UTC m=+732.304447742 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist") pod "speaker-h7xhj" (UID: "2c199acd-2fd3-4b9f-b50c-a266191c4777") : secret "metallb-memberlist" not found Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.332704 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-metrics-certs\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.401675 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs"] Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.470596 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-c6sqk"] Feb 20 08:18:02 crc kubenswrapper[4948]: W0220 08:18:02.483344 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68ee0185_13d4_45c7_9a39_bb474e9314d0.slice/crio-29ad037d86595ebeca4c463306356d180d574a77f75e87dfb0da6a250144e832 WatchSource:0}: Error finding container 29ad037d86595ebeca4c463306356d180d574a77f75e87dfb0da6a250144e832: Status 404 returned error can't find the container with id 29ad037d86595ebeca4c463306356d180d574a77f75e87dfb0da6a250144e832 Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.518020 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.742240 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" event={"ID":"2506d4e2-b434-4fe4-970e-7cd14601677d","Type":"ContainerStarted","Data":"374f01b7e30c3730f2d0fe4500f1a90d86e1a9243240a4cc2a39a2e1093752ac"} Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.743173 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerStarted","Data":"c12d45fc8456567f3dfc5e086c5767ad4130f5148db37efe222ca6aa614aa44f"} Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.753195 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-c6sqk" event={"ID":"68ee0185-13d4-45c7-9a39-bb474e9314d0","Type":"ContainerStarted","Data":"ec977772b2186974da9e5022ec788dee392cac9a842ebd0f652092982d64289e"} Feb 20 08:18:02 crc kubenswrapper[4948]: I0220 08:18:02.753235 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-c6sqk" event={"ID":"68ee0185-13d4-45c7-9a39-bb474e9314d0","Type":"ContainerStarted","Data":"29ad037d86595ebeca4c463306356d180d574a77f75e87dfb0da6a250144e832"} Feb 20 08:18:03 crc kubenswrapper[4948]: I0220 08:18:03.348541 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:03 crc kubenswrapper[4948]: I0220 08:18:03.365008 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2c199acd-2fd3-4b9f-b50c-a266191c4777-memberlist\") pod \"speaker-h7xhj\" (UID: \"2c199acd-2fd3-4b9f-b50c-a266191c4777\") " pod="metallb-system/speaker-h7xhj" Feb 20 08:18:03 crc kubenswrapper[4948]: I0220 08:18:03.495808 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-h7xhj" Feb 20 08:18:03 crc kubenswrapper[4948]: W0220 08:18:03.521411 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c199acd_2fd3_4b9f_b50c_a266191c4777.slice/crio-c448fbd8c6e26ac85c565ff365fb80b2550c88cdb6506d1a672f568a361f309d WatchSource:0}: Error finding container c448fbd8c6e26ac85c565ff365fb80b2550c88cdb6506d1a672f568a361f309d: Status 404 returned error can't find the container with id c448fbd8c6e26ac85c565ff365fb80b2550c88cdb6506d1a672f568a361f309d Feb 20 08:18:03 crc kubenswrapper[4948]: I0220 08:18:03.766274 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-c6sqk" event={"ID":"68ee0185-13d4-45c7-9a39-bb474e9314d0","Type":"ContainerStarted","Data":"05b7f34196b4aa6a1d50d3eab40559177ce6fd571f9dce95d25544ee5991268a"} Feb 20 08:18:03 crc kubenswrapper[4948]: I0220 08:18:03.766339 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:03 crc kubenswrapper[4948]: I0220 08:18:03.767766 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-h7xhj" event={"ID":"2c199acd-2fd3-4b9f-b50c-a266191c4777","Type":"ContainerStarted","Data":"c448fbd8c6e26ac85c565ff365fb80b2550c88cdb6506d1a672f568a361f309d"} Feb 20 08:18:03 crc kubenswrapper[4948]: I0220 08:18:03.791211 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-69bbfbf88f-c6sqk" podStartSLOduration=2.791195977 podStartE2EDuration="2.791195977s" podCreationTimestamp="2026-02-20 08:18:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:18:03.787057433 +0000 UTC m=+732.761552273" watchObservedRunningTime="2026-02-20 08:18:03.791195977 +0000 UTC m=+732.765690797" Feb 20 08:18:04 crc kubenswrapper[4948]: I0220 08:18:04.778467 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-h7xhj" event={"ID":"2c199acd-2fd3-4b9f-b50c-a266191c4777","Type":"ContainerStarted","Data":"bb02cb1ad5d33d591b81172c3773b8ac4a801b8f5d3d3ab8ef4c2647b1e2845c"} Feb 20 08:18:04 crc kubenswrapper[4948]: I0220 08:18:04.779043 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-h7xhj" Feb 20 08:18:04 crc kubenswrapper[4948]: I0220 08:18:04.779066 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-h7xhj" event={"ID":"2c199acd-2fd3-4b9f-b50c-a266191c4777","Type":"ContainerStarted","Data":"8c6273d649cf80ad250c82bc059cd053d11c3997a59f0fdb4f94f58150d57ecc"} Feb 20 08:18:04 crc kubenswrapper[4948]: I0220 08:18:04.813037 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-h7xhj" podStartSLOduration=3.813019675 podStartE2EDuration="3.813019675s" podCreationTimestamp="2026-02-20 08:18:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:18:04.808662146 +0000 UTC m=+733.783156966" watchObservedRunningTime="2026-02-20 08:18:04.813019675 +0000 UTC m=+733.787514495" Feb 20 08:18:08 crc kubenswrapper[4948]: I0220 08:18:08.024958 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:18:08 crc kubenswrapper[4948]: I0220 08:18:08.025344 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:18:09 crc kubenswrapper[4948]: I0220 08:18:09.813037 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" event={"ID":"2506d4e2-b434-4fe4-970e-7cd14601677d","Type":"ContainerStarted","Data":"3e76415ec36aa24893d0adf7e5a2c2da8d861d518e53e80711fde32279aaa0f6"} Feb 20 08:18:09 crc kubenswrapper[4948]: I0220 08:18:09.814226 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:09 crc kubenswrapper[4948]: I0220 08:18:09.816248 4948 generic.go:334] "Generic (PLEG): container finished" podID="0131d05b-619d-43d8-aa38-ad6e9ce52e11" containerID="a9a7882b7d062e23fac330c2ac913b9373cfdb27dbe52941c158234dc75308b7" exitCode=0 Feb 20 08:18:09 crc kubenswrapper[4948]: I0220 08:18:09.816326 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerDied","Data":"a9a7882b7d062e23fac330c2ac913b9373cfdb27dbe52941c158234dc75308b7"} Feb 20 08:18:09 crc kubenswrapper[4948]: I0220 08:18:09.851426 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" podStartSLOduration=2.25378656 podStartE2EDuration="8.851392599s" podCreationTimestamp="2026-02-20 08:18:01 +0000 UTC" firstStartedPulling="2026-02-20 08:18:02.406143267 +0000 UTC m=+731.380638137" lastFinishedPulling="2026-02-20 08:18:09.003749326 +0000 UTC m=+737.978244176" observedRunningTime="2026-02-20 08:18:09.842360474 +0000 UTC m=+738.816855334" watchObservedRunningTime="2026-02-20 08:18:09.851392599 +0000 UTC m=+738.825887459" Feb 20 08:18:10 crc kubenswrapper[4948]: I0220 08:18:10.824798 4948 generic.go:334] "Generic (PLEG): container finished" podID="0131d05b-619d-43d8-aa38-ad6e9ce52e11" containerID="48cfb755827c44a7b77441317598f5844ce40cd1199c0cf03cdcf7ff06e3a166" exitCode=0 Feb 20 08:18:10 crc kubenswrapper[4948]: I0220 08:18:10.824909 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerDied","Data":"48cfb755827c44a7b77441317598f5844ce40cd1199c0cf03cdcf7ff06e3a166"} Feb 20 08:18:11 crc kubenswrapper[4948]: I0220 08:18:11.833176 4948 generic.go:334] "Generic (PLEG): container finished" podID="0131d05b-619d-43d8-aa38-ad6e9ce52e11" containerID="77dbe5ef30ba0d2bfca47dad947f0aa69e25dab5fcc9c3b9cb4e9edb8e2d9e4c" exitCode=0 Feb 20 08:18:11 crc kubenswrapper[4948]: I0220 08:18:11.833281 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerDied","Data":"77dbe5ef30ba0d2bfca47dad947f0aa69e25dab5fcc9c3b9cb4e9edb8e2d9e4c"} Feb 20 08:18:12 crc kubenswrapper[4948]: I0220 08:18:12.031524 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-69bbfbf88f-c6sqk" Feb 20 08:18:12 crc kubenswrapper[4948]: I0220 08:18:12.848404 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerStarted","Data":"30cb9cbaefc7e5bc426ae70da6190592a3e466396c71864cfac194785f1b5e02"} Feb 20 08:18:12 crc kubenswrapper[4948]: I0220 08:18:12.848719 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerStarted","Data":"ed0ce7eb66e6ec5dbf50f538a5187d98f3d7beb4cdb41c88e4cb386935a9b5c3"} Feb 20 08:18:12 crc kubenswrapper[4948]: I0220 08:18:12.848736 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerStarted","Data":"a8ae56528ccc400d0e41549e4070b82a360a0f4a92a643c778435fe75b521045"} Feb 20 08:18:12 crc kubenswrapper[4948]: I0220 08:18:12.848749 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerStarted","Data":"b3b777ca08391e90927b0bb613a5a0899da7dca095fb57f79f9c3ced5d93525d"} Feb 20 08:18:12 crc kubenswrapper[4948]: I0220 08:18:12.848760 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerStarted","Data":"7fa91c580d6df65b49dc14afe2ea7940e9da4a8ffaca47686baa88ca958dab97"} Feb 20 08:18:13 crc kubenswrapper[4948]: I0220 08:18:13.501246 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-h7xhj" Feb 20 08:18:13 crc kubenswrapper[4948]: I0220 08:18:13.860280 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k66nh" event={"ID":"0131d05b-619d-43d8-aa38-ad6e9ce52e11","Type":"ContainerStarted","Data":"5953ccf4702bfdc91094e1f472c02fc87dbf0a81b8c3ec30c9401220c436b465"} Feb 20 08:18:13 crc kubenswrapper[4948]: I0220 08:18:13.860581 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:13 crc kubenswrapper[4948]: I0220 08:18:13.897679 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-k66nh" podStartSLOduration=6.505809565 podStartE2EDuration="12.897657018s" podCreationTimestamp="2026-02-20 08:18:01 +0000 UTC" firstStartedPulling="2026-02-20 08:18:02.636767194 +0000 UTC m=+731.611262014" lastFinishedPulling="2026-02-20 08:18:09.028614617 +0000 UTC m=+738.003109467" observedRunningTime="2026-02-20 08:18:13.892175181 +0000 UTC m=+742.866670051" watchObservedRunningTime="2026-02-20 08:18:13.897657018 +0000 UTC m=+742.872151848" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.192119 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-r6bgs"] Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.193728 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r6bgs" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.197145 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.197594 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.199564 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-7bd79" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.223174 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-r6bgs"] Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.248055 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrlrf\" (UniqueName: \"kubernetes.io/projected/7598c911-bd3c-4772-99f8-668c1bba4495-kube-api-access-jrlrf\") pod \"openstack-operator-index-r6bgs\" (UID: \"7598c911-bd3c-4772-99f8-668c1bba4495\") " pod="openstack-operators/openstack-operator-index-r6bgs" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.349505 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrlrf\" (UniqueName: \"kubernetes.io/projected/7598c911-bd3c-4772-99f8-668c1bba4495-kube-api-access-jrlrf\") pod \"openstack-operator-index-r6bgs\" (UID: \"7598c911-bd3c-4772-99f8-668c1bba4495\") " pod="openstack-operators/openstack-operator-index-r6bgs" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.371435 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrlrf\" (UniqueName: \"kubernetes.io/projected/7598c911-bd3c-4772-99f8-668c1bba4495-kube-api-access-jrlrf\") pod \"openstack-operator-index-r6bgs\" (UID: \"7598c911-bd3c-4772-99f8-668c1bba4495\") " pod="openstack-operators/openstack-operator-index-r6bgs" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.512644 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r6bgs" Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.754344 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-r6bgs"] Feb 20 08:18:16 crc kubenswrapper[4948]: W0220 08:18:16.765506 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7598c911_bd3c_4772_99f8_668c1bba4495.slice/crio-84871b79f61cd574bf003c73f0343d6e3cd10c676f9811ef0b6e00f040669452 WatchSource:0}: Error finding container 84871b79f61cd574bf003c73f0343d6e3cd10c676f9811ef0b6e00f040669452: Status 404 returned error can't find the container with id 84871b79f61cd574bf003c73f0343d6e3cd10c676f9811ef0b6e00f040669452 Feb 20 08:18:16 crc kubenswrapper[4948]: I0220 08:18:16.883799 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r6bgs" event={"ID":"7598c911-bd3c-4772-99f8-668c1bba4495","Type":"ContainerStarted","Data":"84871b79f61cd574bf003c73f0343d6e3cd10c676f9811ef0b6e00f040669452"} Feb 20 08:18:17 crc kubenswrapper[4948]: I0220 08:18:17.519332 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:17 crc kubenswrapper[4948]: I0220 08:18:17.568313 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:19 crc kubenswrapper[4948]: I0220 08:18:19.370315 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-r6bgs"] Feb 20 08:18:19 crc kubenswrapper[4948]: I0220 08:18:19.905770 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r6bgs" event={"ID":"7598c911-bd3c-4772-99f8-668c1bba4495","Type":"ContainerStarted","Data":"5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b"} Feb 20 08:18:19 crc kubenswrapper[4948]: I0220 08:18:19.906026 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-r6bgs" podUID="7598c911-bd3c-4772-99f8-668c1bba4495" containerName="registry-server" containerID="cri-o://5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b" gracePeriod=2 Feb 20 08:18:19 crc kubenswrapper[4948]: I0220 08:18:19.941622 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-r6bgs" podStartSLOduration=1.4406791239999999 podStartE2EDuration="3.941591164s" podCreationTimestamp="2026-02-20 08:18:16 +0000 UTC" firstStartedPulling="2026-02-20 08:18:16.768320043 +0000 UTC m=+745.742814873" lastFinishedPulling="2026-02-20 08:18:19.269232103 +0000 UTC m=+748.243726913" observedRunningTime="2026-02-20 08:18:19.931621275 +0000 UTC m=+748.906116155" watchObservedRunningTime="2026-02-20 08:18:19.941591164 +0000 UTC m=+748.916086024" Feb 20 08:18:19 crc kubenswrapper[4948]: I0220 08:18:19.971005 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-9z4jj"] Feb 20 08:18:19 crc kubenswrapper[4948]: I0220 08:18:19.971852 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:19 crc kubenswrapper[4948]: I0220 08:18:19.988182 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-9z4jj"] Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.026436 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrwkr\" (UniqueName: \"kubernetes.io/projected/64872326-fb95-4bc1-a6c5-0b34242883f9-kube-api-access-lrwkr\") pod \"openstack-operator-index-9z4jj\" (UID: \"64872326-fb95-4bc1-a6c5-0b34242883f9\") " pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.127770 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrwkr\" (UniqueName: \"kubernetes.io/projected/64872326-fb95-4bc1-a6c5-0b34242883f9-kube-api-access-lrwkr\") pod \"openstack-operator-index-9z4jj\" (UID: \"64872326-fb95-4bc1-a6c5-0b34242883f9\") " pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.166806 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrwkr\" (UniqueName: \"kubernetes.io/projected/64872326-fb95-4bc1-a6c5-0b34242883f9-kube-api-access-lrwkr\") pod \"openstack-operator-index-9z4jj\" (UID: \"64872326-fb95-4bc1-a6c5-0b34242883f9\") " pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.283370 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r6bgs" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.363336 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.449683 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrlrf\" (UniqueName: \"kubernetes.io/projected/7598c911-bd3c-4772-99f8-668c1bba4495-kube-api-access-jrlrf\") pod \"7598c911-bd3c-4772-99f8-668c1bba4495\" (UID: \"7598c911-bd3c-4772-99f8-668c1bba4495\") " Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.454133 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7598c911-bd3c-4772-99f8-668c1bba4495-kube-api-access-jrlrf" (OuterVolumeSpecName: "kube-api-access-jrlrf") pod "7598c911-bd3c-4772-99f8-668c1bba4495" (UID: "7598c911-bd3c-4772-99f8-668c1bba4495"). InnerVolumeSpecName "kube-api-access-jrlrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.553534 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrlrf\" (UniqueName: \"kubernetes.io/projected/7598c911-bd3c-4772-99f8-668c1bba4495-kube-api-access-jrlrf\") on node \"crc\" DevicePath \"\"" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.759478 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-9z4jj"] Feb 20 08:18:20 crc kubenswrapper[4948]: W0220 08:18:20.769785 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64872326_fb95_4bc1_a6c5_0b34242883f9.slice/crio-bd0dff1cb53c89e153782945c052c068ebbad59155c54bd5132f267675f97e82 WatchSource:0}: Error finding container bd0dff1cb53c89e153782945c052c068ebbad59155c54bd5132f267675f97e82: Status 404 returned error can't find the container with id bd0dff1cb53c89e153782945c052c068ebbad59155c54bd5132f267675f97e82 Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.915366 4948 generic.go:334] "Generic (PLEG): container finished" podID="7598c911-bd3c-4772-99f8-668c1bba4495" containerID="5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b" exitCode=0 Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.915502 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r6bgs" event={"ID":"7598c911-bd3c-4772-99f8-668c1bba4495","Type":"ContainerDied","Data":"5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b"} Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.915512 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r6bgs" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.915819 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r6bgs" event={"ID":"7598c911-bd3c-4772-99f8-668c1bba4495","Type":"ContainerDied","Data":"84871b79f61cd574bf003c73f0343d6e3cd10c676f9811ef0b6e00f040669452"} Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.915862 4948 scope.go:117] "RemoveContainer" containerID="5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.917201 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-9z4jj" event={"ID":"64872326-fb95-4bc1-a6c5-0b34242883f9","Type":"ContainerStarted","Data":"bd0dff1cb53c89e153782945c052c068ebbad59155c54bd5132f267675f97e82"} Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.936347 4948 scope.go:117] "RemoveContainer" containerID="5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b" Feb 20 08:18:20 crc kubenswrapper[4948]: E0220 08:18:20.936925 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b\": container with ID starting with 5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b not found: ID does not exist" containerID="5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.936958 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b"} err="failed to get container status \"5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b\": rpc error: code = NotFound desc = could not find container \"5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b\": container with ID starting with 5bda4722a9fc484672b67eadaa90074e99796b98e5bc9ff3ced68ecd54a8925b not found: ID does not exist" Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.983420 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-r6bgs"] Feb 20 08:18:20 crc kubenswrapper[4948]: I0220 08:18:20.988521 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-r6bgs"] Feb 20 08:18:21 crc kubenswrapper[4948]: I0220 08:18:21.733151 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7598c911-bd3c-4772-99f8-668c1bba4495" path="/var/lib/kubelet/pods/7598c911-bd3c-4772-99f8-668c1bba4495/volumes" Feb 20 08:18:21 crc kubenswrapper[4948]: I0220 08:18:21.928363 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-9z4jj" event={"ID":"64872326-fb95-4bc1-a6c5-0b34242883f9","Type":"ContainerStarted","Data":"9e917cc57712b696be945e95695645211dadf6c831fb65d4dd1e15d4de7f4f41"} Feb 20 08:18:21 crc kubenswrapper[4948]: I0220 08:18:21.945013 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7zsxs" Feb 20 08:18:21 crc kubenswrapper[4948]: I0220 08:18:21.950586 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-9z4jj" podStartSLOduration=2.896904293 podStartE2EDuration="2.950560525s" podCreationTimestamp="2026-02-20 08:18:19 +0000 UTC" firstStartedPulling="2026-02-20 08:18:20.779283199 +0000 UTC m=+749.753778049" lastFinishedPulling="2026-02-20 08:18:20.832939461 +0000 UTC m=+749.807434281" observedRunningTime="2026-02-20 08:18:21.948014641 +0000 UTC m=+750.922509491" watchObservedRunningTime="2026-02-20 08:18:21.950560525 +0000 UTC m=+750.925055385" Feb 20 08:18:22 crc kubenswrapper[4948]: I0220 08:18:22.523318 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-k66nh" Feb 20 08:18:30 crc kubenswrapper[4948]: I0220 08:18:30.364542 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:30 crc kubenswrapper[4948]: I0220 08:18:30.365202 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:30 crc kubenswrapper[4948]: I0220 08:18:30.406163 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:31 crc kubenswrapper[4948]: I0220 08:18:31.034161 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-9z4jj" Feb 20 08:18:35 crc kubenswrapper[4948]: I0220 08:18:35.217688 4948 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.025472 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.025883 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.025942 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.026662 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b5405a03f834a339ac758e6efb1e31a8c094b3f6b973ea8540eefc716ba7bf7f"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.026729 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://b5405a03f834a339ac758e6efb1e31a8c094b3f6b973ea8540eefc716ba7bf7f" gracePeriod=600 Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.967366 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6"] Feb 20 08:18:38 crc kubenswrapper[4948]: E0220 08:18:38.968323 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7598c911-bd3c-4772-99f8-668c1bba4495" containerName="registry-server" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.968724 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="7598c911-bd3c-4772-99f8-668c1bba4495" containerName="registry-server" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.968890 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="7598c911-bd3c-4772-99f8-668c1bba4495" containerName="registry-server" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.969908 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.972124 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-z6fxt" Feb 20 08:18:38 crc kubenswrapper[4948]: I0220 08:18:38.979347 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6"] Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.030182 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-util\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.030254 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr5qj\" (UniqueName: \"kubernetes.io/projected/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-kube-api-access-tr5qj\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.030434 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-bundle\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.067387 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="b5405a03f834a339ac758e6efb1e31a8c094b3f6b973ea8540eefc716ba7bf7f" exitCode=0 Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.067416 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"b5405a03f834a339ac758e6efb1e31a8c094b3f6b973ea8540eefc716ba7bf7f"} Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.067762 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"df8e056c06ac241c46a927c4e3f5a28f39d042bf13a8c672499d528786dbbbb0"} Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.067859 4948 scope.go:117] "RemoveContainer" containerID="a2c247070dfb4d0488751bb46e5543e834b3853c983d98f00e419c6aabfdeb7c" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.131724 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-bundle\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.131889 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-util\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.132092 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr5qj\" (UniqueName: \"kubernetes.io/projected/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-kube-api-access-tr5qj\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.132547 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-bundle\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.133060 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-util\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.163697 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr5qj\" (UniqueName: \"kubernetes.io/projected/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-kube-api-access-tr5qj\") pod \"b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.291042 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:39 crc kubenswrapper[4948]: I0220 08:18:39.476137 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6"] Feb 20 08:18:39 crc kubenswrapper[4948]: W0220 08:18:39.489468 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1b99eb4_2ab2_4755_92ed_a5226d91bfaf.slice/crio-56819cc9445f12d4a1429b2a438986922d555b3060e1030ae43bca1e151aca92 WatchSource:0}: Error finding container 56819cc9445f12d4a1429b2a438986922d555b3060e1030ae43bca1e151aca92: Status 404 returned error can't find the container with id 56819cc9445f12d4a1429b2a438986922d555b3060e1030ae43bca1e151aca92 Feb 20 08:18:40 crc kubenswrapper[4948]: I0220 08:18:40.076605 4948 generic.go:334] "Generic (PLEG): container finished" podID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerID="33c724572b1849d1e782827fa80b211bb380e77669c568898c85b28655569808" exitCode=0 Feb 20 08:18:40 crc kubenswrapper[4948]: I0220 08:18:40.076656 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" event={"ID":"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf","Type":"ContainerDied","Data":"33c724572b1849d1e782827fa80b211bb380e77669c568898c85b28655569808"} Feb 20 08:18:40 crc kubenswrapper[4948]: I0220 08:18:40.077093 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" event={"ID":"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf","Type":"ContainerStarted","Data":"56819cc9445f12d4a1429b2a438986922d555b3060e1030ae43bca1e151aca92"} Feb 20 08:18:41 crc kubenswrapper[4948]: I0220 08:18:41.091665 4948 generic.go:334] "Generic (PLEG): container finished" podID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerID="82d5bea9ab2ae66c0e5fba37f18d49d29e12c46a302a57e13775cc41280a1d46" exitCode=0 Feb 20 08:18:41 crc kubenswrapper[4948]: I0220 08:18:41.091815 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" event={"ID":"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf","Type":"ContainerDied","Data":"82d5bea9ab2ae66c0e5fba37f18d49d29e12c46a302a57e13775cc41280a1d46"} Feb 20 08:18:42 crc kubenswrapper[4948]: I0220 08:18:42.105959 4948 generic.go:334] "Generic (PLEG): container finished" podID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerID="0bac1a873808460bd3bbd795e40f16ef73fcda97d242286fa6c592d23ce18f0e" exitCode=0 Feb 20 08:18:42 crc kubenswrapper[4948]: I0220 08:18:42.106088 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" event={"ID":"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf","Type":"ContainerDied","Data":"0bac1a873808460bd3bbd795e40f16ef73fcda97d242286fa6c592d23ce18f0e"} Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.508560 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.609519 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-util\") pod \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.609930 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-bundle\") pod \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.609974 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tr5qj\" (UniqueName: \"kubernetes.io/projected/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-kube-api-access-tr5qj\") pod \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\" (UID: \"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf\") " Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.610633 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-bundle" (OuterVolumeSpecName: "bundle") pod "a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" (UID: "a1b99eb4-2ab2-4755-92ed-a5226d91bfaf"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.616680 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-kube-api-access-tr5qj" (OuterVolumeSpecName: "kube-api-access-tr5qj") pod "a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" (UID: "a1b99eb4-2ab2-4755-92ed-a5226d91bfaf"). InnerVolumeSpecName "kube-api-access-tr5qj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.629167 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-util" (OuterVolumeSpecName: "util") pod "a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" (UID: "a1b99eb4-2ab2-4755-92ed-a5226d91bfaf"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.712145 4948 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-util\") on node \"crc\" DevicePath \"\"" Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.712198 4948 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:18:43 crc kubenswrapper[4948]: I0220 08:18:43.712219 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tr5qj\" (UniqueName: \"kubernetes.io/projected/a1b99eb4-2ab2-4755-92ed-a5226d91bfaf-kube-api-access-tr5qj\") on node \"crc\" DevicePath \"\"" Feb 20 08:18:44 crc kubenswrapper[4948]: I0220 08:18:44.127020 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" event={"ID":"a1b99eb4-2ab2-4755-92ed-a5226d91bfaf","Type":"ContainerDied","Data":"56819cc9445f12d4a1429b2a438986922d555b3060e1030ae43bca1e151aca92"} Feb 20 08:18:44 crc kubenswrapper[4948]: I0220 08:18:44.127079 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56819cc9445f12d4a1429b2a438986922d555b3060e1030ae43bca1e151aca92" Feb 20 08:18:44 crc kubenswrapper[4948]: I0220 08:18:44.127115 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.239199 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n"] Feb 20 08:18:51 crc kubenswrapper[4948]: E0220 08:18:51.240111 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerName="util" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.240128 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerName="util" Feb 20 08:18:51 crc kubenswrapper[4948]: E0220 08:18:51.240150 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerName="extract" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.240157 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerName="extract" Feb 20 08:18:51 crc kubenswrapper[4948]: E0220 08:18:51.240172 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerName="pull" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.240196 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerName="pull" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.240321 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1b99eb4-2ab2-4755-92ed-a5226d91bfaf" containerName="extract" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.240735 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.243771 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-9hmd7" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.323234 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n"] Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.323780 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4899m\" (UniqueName: \"kubernetes.io/projected/9465a7a8-c08e-44c9-a76d-3878b0bff3af-kube-api-access-4899m\") pod \"openstack-operator-controller-init-59c446d8d6-4cs2n\" (UID: \"9465a7a8-c08e-44c9-a76d-3878b0bff3af\") " pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.424667 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4899m\" (UniqueName: \"kubernetes.io/projected/9465a7a8-c08e-44c9-a76d-3878b0bff3af-kube-api-access-4899m\") pod \"openstack-operator-controller-init-59c446d8d6-4cs2n\" (UID: \"9465a7a8-c08e-44c9-a76d-3878b0bff3af\") " pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.444817 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4899m\" (UniqueName: \"kubernetes.io/projected/9465a7a8-c08e-44c9-a76d-3878b0bff3af-kube-api-access-4899m\") pod \"openstack-operator-controller-init-59c446d8d6-4cs2n\" (UID: \"9465a7a8-c08e-44c9-a76d-3878b0bff3af\") " pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.574710 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" Feb 20 08:18:51 crc kubenswrapper[4948]: I0220 08:18:51.789039 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n"] Feb 20 08:18:51 crc kubenswrapper[4948]: W0220 08:18:51.798733 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9465a7a8_c08e_44c9_a76d_3878b0bff3af.slice/crio-151160100b0cd6df4f324891edcdc7d3493e3f9a61b069450e63c6482f0a924c WatchSource:0}: Error finding container 151160100b0cd6df4f324891edcdc7d3493e3f9a61b069450e63c6482f0a924c: Status 404 returned error can't find the container with id 151160100b0cd6df4f324891edcdc7d3493e3f9a61b069450e63c6482f0a924c Feb 20 08:18:52 crc kubenswrapper[4948]: I0220 08:18:52.207891 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" event={"ID":"9465a7a8-c08e-44c9-a76d-3878b0bff3af","Type":"ContainerStarted","Data":"151160100b0cd6df4f324891edcdc7d3493e3f9a61b069450e63c6482f0a924c"} Feb 20 08:18:56 crc kubenswrapper[4948]: I0220 08:18:56.247370 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" event={"ID":"9465a7a8-c08e-44c9-a76d-3878b0bff3af","Type":"ContainerStarted","Data":"a4b82b549f81b88e36ef7dc8eb38618f1e90051ff837cadf4dd63fa70291dd3f"} Feb 20 08:18:56 crc kubenswrapper[4948]: I0220 08:18:56.247869 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" Feb 20 08:18:56 crc kubenswrapper[4948]: I0220 08:18:56.279504 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" podStartSLOduration=1.715627904 podStartE2EDuration="5.279480194s" podCreationTimestamp="2026-02-20 08:18:51 +0000 UTC" firstStartedPulling="2026-02-20 08:18:51.800480279 +0000 UTC m=+780.774975109" lastFinishedPulling="2026-02-20 08:18:55.364332569 +0000 UTC m=+784.338827399" observedRunningTime="2026-02-20 08:18:56.279167687 +0000 UTC m=+785.253662517" watchObservedRunningTime="2026-02-20 08:18:56.279480194 +0000 UTC m=+785.253975054" Feb 20 08:19:01 crc kubenswrapper[4948]: I0220 08:19:01.577668 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-59c446d8d6-4cs2n" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.536276 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.537791 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.540146 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-5fbgl" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.543379 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.544430 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.546235 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-vspgb" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.550991 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.557499 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.566070 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.567401 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.575625 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-r8cz9" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.613927 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.631068 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.631790 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.633191 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-qt8sb" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.642752 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-xc74s"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.643702 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.650541 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zgxs\" (UniqueName: \"kubernetes.io/projected/586113f2-38b6-4bd1-8adb-3c155bb35ba9-kube-api-access-7zgxs\") pod \"cinder-operator-controller-manager-5d946d989d-c6xqd\" (UID: \"586113f2-38b6-4bd1-8adb-3c155bb35ba9\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.650603 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwh5b\" (UniqueName: \"kubernetes.io/projected/b8e12d0c-9564-4bf2-ac61-b22d2fbdf855-kube-api-access-gwh5b\") pod \"barbican-operator-controller-manager-868647ff47-4h9v5\" (UID: \"b8e12d0c-9564-4bf2-ac61-b22d2fbdf855\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.650615 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-6m82q" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.650635 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9zrw\" (UniqueName: \"kubernetes.io/projected/98a1aa06-948b-4034-bc07-7e546e341a8f-kube-api-access-d9zrw\") pod \"designate-operator-controller-manager-6d8bf5c495-kjpkr\" (UID: \"98a1aa06-948b-4034-bc07-7e546e341a8f\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.651039 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-xc74s"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.657753 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.659815 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.660894 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.673586 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-57g8d" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.679093 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.682072 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.682769 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.684642 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.687183 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-nxfl4" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.721681 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.722606 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.729417 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-5vbfv" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.742183 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.742224 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.753765 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr4jf\" (UniqueName: \"kubernetes.io/projected/2d303d1a-1062-4bf6-be68-1bd6d1a3228f-kube-api-access-gr4jf\") pod \"heat-operator-controller-manager-69f49c598c-6cr5p\" (UID: \"2d303d1a-1062-4bf6-be68-1bd6d1a3228f\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.753820 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rd7r\" (UniqueName: \"kubernetes.io/projected/4a663826-d4b5-4ed3-8270-099b003390b4-kube-api-access-8rd7r\") pod \"horizon-operator-controller-manager-5b9b8895d5-bkl49\" (UID: \"4a663826-d4b5-4ed3-8270-099b003390b4\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.753892 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht8br\" (UniqueName: \"kubernetes.io/projected/347cdd7c-e5e8-49b1-a4f1-687b8a06b250-kube-api-access-ht8br\") pod \"glance-operator-controller-manager-77987464f4-xc74s\" (UID: \"347cdd7c-e5e8-49b1-a4f1-687b8a06b250\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.753944 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zgxs\" (UniqueName: \"kubernetes.io/projected/586113f2-38b6-4bd1-8adb-3c155bb35ba9-kube-api-access-7zgxs\") pod \"cinder-operator-controller-manager-5d946d989d-c6xqd\" (UID: \"586113f2-38b6-4bd1-8adb-3c155bb35ba9\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.754009 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwh5b\" (UniqueName: \"kubernetes.io/projected/b8e12d0c-9564-4bf2-ac61-b22d2fbdf855-kube-api-access-gwh5b\") pod \"barbican-operator-controller-manager-868647ff47-4h9v5\" (UID: \"b8e12d0c-9564-4bf2-ac61-b22d2fbdf855\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.754042 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9zrw\" (UniqueName: \"kubernetes.io/projected/98a1aa06-948b-4034-bc07-7e546e341a8f-kube-api-access-d9zrw\") pod \"designate-operator-controller-manager-6d8bf5c495-kjpkr\" (UID: \"98a1aa06-948b-4034-bc07-7e546e341a8f\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.756741 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.757781 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.759640 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-b44t8" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.765651 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.766650 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.771832 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-qhs6f" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.792178 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9zrw\" (UniqueName: \"kubernetes.io/projected/98a1aa06-948b-4034-bc07-7e546e341a8f-kube-api-access-d9zrw\") pod \"designate-operator-controller-manager-6d8bf5c495-kjpkr\" (UID: \"98a1aa06-948b-4034-bc07-7e546e341a8f\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.793334 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.800715 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwh5b\" (UniqueName: \"kubernetes.io/projected/b8e12d0c-9564-4bf2-ac61-b22d2fbdf855-kube-api-access-gwh5b\") pod \"barbican-operator-controller-manager-868647ff47-4h9v5\" (UID: \"b8e12d0c-9564-4bf2-ac61-b22d2fbdf855\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.805158 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zgxs\" (UniqueName: \"kubernetes.io/projected/586113f2-38b6-4bd1-8adb-3c155bb35ba9-kube-api-access-7zgxs\") pod \"cinder-operator-controller-manager-5d946d989d-c6xqd\" (UID: \"586113f2-38b6-4bd1-8adb-3c155bb35ba9\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.808882 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.809784 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.814872 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-2njl7" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.817363 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.856574 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht8br\" (UniqueName: \"kubernetes.io/projected/347cdd7c-e5e8-49b1-a4f1-687b8a06b250-kube-api-access-ht8br\") pod \"glance-operator-controller-manager-77987464f4-xc74s\" (UID: \"347cdd7c-e5e8-49b1-a4f1-687b8a06b250\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.856651 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbh26\" (UniqueName: \"kubernetes.io/projected/93b93501-acfe-4274-9a22-ca644b1d11d3-kube-api-access-mbh26\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.856681 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvjkg\" (UniqueName: \"kubernetes.io/projected/105a118c-121d-4582-960f-1da9957980cb-kube-api-access-hvjkg\") pod \"ironic-operator-controller-manager-554564d7fc-5fqzw\" (UID: \"105a118c-121d-4582-960f-1da9957980cb\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.856741 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.856779 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr4jf\" (UniqueName: \"kubernetes.io/projected/2d303d1a-1062-4bf6-be68-1bd6d1a3228f-kube-api-access-gr4jf\") pod \"heat-operator-controller-manager-69f49c598c-6cr5p\" (UID: \"2d303d1a-1062-4bf6-be68-1bd6d1a3228f\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.856801 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rd7r\" (UniqueName: \"kubernetes.io/projected/4a663826-d4b5-4ed3-8270-099b003390b4-kube-api-access-8rd7r\") pod \"horizon-operator-controller-manager-5b9b8895d5-bkl49\" (UID: \"4a663826-d4b5-4ed3-8270-099b003390b4\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.856897 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28rlg\" (UniqueName: \"kubernetes.io/projected/3fbbd11d-612d-479a-a34f-505d995a4871-kube-api-access-28rlg\") pod \"keystone-operator-controller-manager-b4d948c87-ttkfm\" (UID: \"3fbbd11d-612d-479a-a34f-505d995a4871\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.856984 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pphn\" (UniqueName: \"kubernetes.io/projected/f762ec0d-e09c-4baf-9540-f4ac61ce7234-kube-api-access-6pphn\") pod \"manila-operator-controller-manager-54f6768c69-lj2sz\" (UID: \"f762ec0d-e09c-4baf-9540-f4ac61ce7234\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.868021 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.876079 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.879456 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.880439 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht8br\" (UniqueName: \"kubernetes.io/projected/347cdd7c-e5e8-49b1-a4f1-687b8a06b250-kube-api-access-ht8br\") pod \"glance-operator-controller-manager-77987464f4-xc74s\" (UID: \"347cdd7c-e5e8-49b1-a4f1-687b8a06b250\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.880723 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rd7r\" (UniqueName: \"kubernetes.io/projected/4a663826-d4b5-4ed3-8270-099b003390b4-kube-api-access-8rd7r\") pod \"horizon-operator-controller-manager-5b9b8895d5-bkl49\" (UID: \"4a663826-d4b5-4ed3-8270-099b003390b4\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.880835 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr4jf\" (UniqueName: \"kubernetes.io/projected/2d303d1a-1062-4bf6-be68-1bd6d1a3228f-kube-api-access-gr4jf\") pod \"heat-operator-controller-manager-69f49c598c-6cr5p\" (UID: \"2d303d1a-1062-4bf6-be68-1bd6d1a3228f\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.882217 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-v67vx" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.882592 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.890464 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.891396 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.893107 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-72llz" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.896508 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.897391 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.903932 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.910225 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.911772 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.918320 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-q8jj9" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.929093 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.933980 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.940572 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.942071 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.944048 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.944420 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-c7nf5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.953448 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.954355 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.956958 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-nhpm5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.957630 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz"] Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.957680 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.957727 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28rlg\" (UniqueName: \"kubernetes.io/projected/3fbbd11d-612d-479a-a34f-505d995a4871-kube-api-access-28rlg\") pod \"keystone-operator-controller-manager-b4d948c87-ttkfm\" (UID: \"3fbbd11d-612d-479a-a34f-505d995a4871\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.957746 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pphn\" (UniqueName: \"kubernetes.io/projected/f762ec0d-e09c-4baf-9540-f4ac61ce7234-kube-api-access-6pphn\") pod \"manila-operator-controller-manager-54f6768c69-lj2sz\" (UID: \"f762ec0d-e09c-4baf-9540-f4ac61ce7234\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.957773 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjkn5\" (UniqueName: \"kubernetes.io/projected/6e8e8b87-5a00-486f-a00c-1450c109f3b2-kube-api-access-vjkn5\") pod \"mariadb-operator-controller-manager-6994f66f48-zcccd\" (UID: \"6e8e8b87-5a00-486f-a00c-1450c109f3b2\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.957809 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbh26\" (UniqueName: \"kubernetes.io/projected/93b93501-acfe-4274-9a22-ca644b1d11d3-kube-api-access-mbh26\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.957835 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvjkg\" (UniqueName: \"kubernetes.io/projected/105a118c-121d-4582-960f-1da9957980cb-kube-api-access-hvjkg\") pod \"ironic-operator-controller-manager-554564d7fc-5fqzw\" (UID: \"105a118c-121d-4582-960f-1da9957980cb\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" Feb 20 08:19:39 crc kubenswrapper[4948]: E0220 08:19:39.958105 4948 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:39 crc kubenswrapper[4948]: E0220 08:19:39.958154 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert podName:93b93501-acfe-4274-9a22-ca644b1d11d3 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:40.458137939 +0000 UTC m=+829.432632759 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert") pod "infra-operator-controller-manager-79d975b745-jrlw5" (UID: "93b93501-acfe-4274-9a22-ca644b1d11d3") : secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.958262 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.964619 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-vvqgl" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.966314 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.981316 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.985684 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvjkg\" (UniqueName: \"kubernetes.io/projected/105a118c-121d-4582-960f-1da9957980cb-kube-api-access-hvjkg\") pod \"ironic-operator-controller-manager-554564d7fc-5fqzw\" (UID: \"105a118c-121d-4582-960f-1da9957980cb\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.985835 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbh26\" (UniqueName: \"kubernetes.io/projected/93b93501-acfe-4274-9a22-ca644b1d11d3-kube-api-access-mbh26\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.991651 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28rlg\" (UniqueName: \"kubernetes.io/projected/3fbbd11d-612d-479a-a34f-505d995a4871-kube-api-access-28rlg\") pod \"keystone-operator-controller-manager-b4d948c87-ttkfm\" (UID: \"3fbbd11d-612d-479a-a34f-505d995a4871\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" Feb 20 08:19:39 crc kubenswrapper[4948]: I0220 08:19:39.995537 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pphn\" (UniqueName: \"kubernetes.io/projected/f762ec0d-e09c-4baf-9540-f4ac61ce7234-kube-api-access-6pphn\") pod \"manila-operator-controller-manager-54f6768c69-lj2sz\" (UID: \"f762ec0d-e09c-4baf-9540-f4ac61ce7234\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.007128 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.027632 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.027672 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.035153 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.045458 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-5lfts"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.048508 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.054469 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-wmrt2" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.054672 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-5lfts"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.054946 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.059006 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x5rb\" (UniqueName: \"kubernetes.io/projected/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-kube-api-access-6x5rb\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.059060 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgpnv\" (UniqueName: \"kubernetes.io/projected/ee33bd02-a3ce-415d-9d2d-fefd383e9810-kube-api-access-fgpnv\") pod \"octavia-operator-controller-manager-69f8888797-hl5w7\" (UID: \"ee33bd02-a3ce-415d-9d2d-fefd383e9810\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.059106 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw4bv\" (UniqueName: \"kubernetes.io/projected/88142137-864d-4660-a688-a7dcc503851b-kube-api-access-xw4bv\") pod \"neutron-operator-controller-manager-64ddbf8bb-jcnxd\" (UID: \"88142137-864d-4660-a688-a7dcc503851b\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.059138 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9srrq\" (UniqueName: \"kubernetes.io/projected/16933747-642c-45ff-9f98-9321c633826a-kube-api-access-9srrq\") pod \"placement-operator-controller-manager-8497b45c89-znrkz\" (UID: \"16933747-642c-45ff-9f98-9321c633826a\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.059162 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd2fq\" (UniqueName: \"kubernetes.io/projected/7c7bb531-7900-4cc6-9d9b-bae52dabc59d-kube-api-access-kd2fq\") pod \"nova-operator-controller-manager-567668f5cf-jxwsx\" (UID: \"7c7bb531-7900-4cc6-9d9b-bae52dabc59d\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.059221 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.059255 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2csx\" (UniqueName: \"kubernetes.io/projected/3ea675de-d1b0-4880-9652-eb066f6b0fb7-kube-api-access-q2csx\") pod \"ovn-operator-controller-manager-d44cf6b75-dkkq9\" (UID: \"3ea675de-d1b0-4880-9652-eb066f6b0fb7\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.059311 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjkn5\" (UniqueName: \"kubernetes.io/projected/6e8e8b87-5a00-486f-a00c-1450c109f3b2-kube-api-access-vjkn5\") pod \"mariadb-operator-controller-manager-6994f66f48-zcccd\" (UID: \"6e8e8b87-5a00-486f-a00c-1450c109f3b2\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.083394 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.084721 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjkn5\" (UniqueName: \"kubernetes.io/projected/6e8e8b87-5a00-486f-a00c-1450c109f3b2-kube-api-access-vjkn5\") pod \"mariadb-operator-controller-manager-6994f66f48-zcccd\" (UID: \"6e8e8b87-5a00-486f-a00c-1450c109f3b2\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.091289 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.093713 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.095547 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.095760 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-8psrm" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.154339 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.156990 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-f7cxn"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.157842 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.160674 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.160718 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2csx\" (UniqueName: \"kubernetes.io/projected/3ea675de-d1b0-4880-9652-eb066f6b0fb7-kube-api-access-q2csx\") pod \"ovn-operator-controller-manager-d44cf6b75-dkkq9\" (UID: \"3ea675de-d1b0-4880-9652-eb066f6b0fb7\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.160777 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxmkg\" (UniqueName: \"kubernetes.io/projected/d65e0993-11a0-4e81-963f-eeb9dcb92536-kube-api-access-xxmkg\") pod \"swift-operator-controller-manager-68f46476f-5lfts\" (UID: \"d65e0993-11a0-4e81-963f-eeb9dcb92536\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.160805 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x5rb\" (UniqueName: \"kubernetes.io/projected/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-kube-api-access-6x5rb\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.160831 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgpnv\" (UniqueName: \"kubernetes.io/projected/ee33bd02-a3ce-415d-9d2d-fefd383e9810-kube-api-access-fgpnv\") pod \"octavia-operator-controller-manager-69f8888797-hl5w7\" (UID: \"ee33bd02-a3ce-415d-9d2d-fefd383e9810\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.160851 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw4bv\" (UniqueName: \"kubernetes.io/projected/88142137-864d-4660-a688-a7dcc503851b-kube-api-access-xw4bv\") pod \"neutron-operator-controller-manager-64ddbf8bb-jcnxd\" (UID: \"88142137-864d-4660-a688-a7dcc503851b\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.160871 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9srrq\" (UniqueName: \"kubernetes.io/projected/16933747-642c-45ff-9f98-9321c633826a-kube-api-access-9srrq\") pod \"placement-operator-controller-manager-8497b45c89-znrkz\" (UID: \"16933747-642c-45ff-9f98-9321c633826a\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.160889 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kd2fq\" (UniqueName: \"kubernetes.io/projected/7c7bb531-7900-4cc6-9d9b-bae52dabc59d-kube-api-access-kd2fq\") pod \"nova-operator-controller-manager-567668f5cf-jxwsx\" (UID: \"7c7bb531-7900-4cc6-9d9b-bae52dabc59d\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" Feb 20 08:19:40 crc kubenswrapper[4948]: E0220 08:19:40.161266 4948 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:40 crc kubenswrapper[4948]: E0220 08:19:40.161313 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert podName:b8b07685-095c-4bbd-a30d-57ca59d7cbdc nodeName:}" failed. No retries permitted until 2026-02-20 08:19:40.66129961 +0000 UTC m=+829.635794430 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" (UID: "b8b07685-095c-4bbd-a30d-57ca59d7cbdc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.161375 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-6zx75" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.176309 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.179455 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kd2fq\" (UniqueName: \"kubernetes.io/projected/7c7bb531-7900-4cc6-9d9b-bae52dabc59d-kube-api-access-kd2fq\") pod \"nova-operator-controller-manager-567668f5cf-jxwsx\" (UID: \"7c7bb531-7900-4cc6-9d9b-bae52dabc59d\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.182540 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2csx\" (UniqueName: \"kubernetes.io/projected/3ea675de-d1b0-4880-9652-eb066f6b0fb7-kube-api-access-q2csx\") pod \"ovn-operator-controller-manager-d44cf6b75-dkkq9\" (UID: \"3ea675de-d1b0-4880-9652-eb066f6b0fb7\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.186959 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9srrq\" (UniqueName: \"kubernetes.io/projected/16933747-642c-45ff-9f98-9321c633826a-kube-api-access-9srrq\") pod \"placement-operator-controller-manager-8497b45c89-znrkz\" (UID: \"16933747-642c-45ff-9f98-9321c633826a\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.188745 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgpnv\" (UniqueName: \"kubernetes.io/projected/ee33bd02-a3ce-415d-9d2d-fefd383e9810-kube-api-access-fgpnv\") pod \"octavia-operator-controller-manager-69f8888797-hl5w7\" (UID: \"ee33bd02-a3ce-415d-9d2d-fefd383e9810\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.188833 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-f7cxn"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.197171 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw4bv\" (UniqueName: \"kubernetes.io/projected/88142137-864d-4660-a688-a7dcc503851b-kube-api-access-xw4bv\") pod \"neutron-operator-controller-manager-64ddbf8bb-jcnxd\" (UID: \"88142137-864d-4660-a688-a7dcc503851b\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.201747 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x5rb\" (UniqueName: \"kubernetes.io/projected/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-kube-api-access-6x5rb\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.237755 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.238960 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.242882 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-qsgww" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.251071 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.259614 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.261448 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxmkg\" (UniqueName: \"kubernetes.io/projected/d65e0993-11a0-4e81-963f-eeb9dcb92536-kube-api-access-xxmkg\") pod \"swift-operator-controller-manager-68f46476f-5lfts\" (UID: \"d65e0993-11a0-4e81-963f-eeb9dcb92536\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.261482 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rjvk\" (UniqueName: \"kubernetes.io/projected/0f98bc58-d0e3-405b-88fd-d8bd65f415a4-kube-api-access-8rjvk\") pod \"test-operator-controller-manager-7866795846-f7cxn\" (UID: \"0f98bc58-d0e3-405b-88fd-d8bd65f415a4\") " pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.261511 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2f9x\" (UniqueName: \"kubernetes.io/projected/0f0cbb65-3c8e-41e6-8059-34e121de0821-kube-api-access-w2f9x\") pod \"telemetry-operator-controller-manager-7f45b4ff68-5rrd9\" (UID: \"0f0cbb65-3c8e-41e6-8059-34e121de0821\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.283063 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxmkg\" (UniqueName: \"kubernetes.io/projected/d65e0993-11a0-4e81-963f-eeb9dcb92536-kube-api-access-xxmkg\") pod \"swift-operator-controller-manager-68f46476f-5lfts\" (UID: \"d65e0993-11a0-4e81-963f-eeb9dcb92536\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.284308 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.286747 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.288394 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.291300 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.292184 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.293861 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-5sbbv" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.295195 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.302982 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.335891 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.341569 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.355680 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-9rn66" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.362906 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rjvk\" (UniqueName: \"kubernetes.io/projected/0f98bc58-d0e3-405b-88fd-d8bd65f415a4-kube-api-access-8rjvk\") pod \"test-operator-controller-manager-7866795846-f7cxn\" (UID: \"0f98bc58-d0e3-405b-88fd-d8bd65f415a4\") " pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.364092 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scvsh\" (UniqueName: \"kubernetes.io/projected/f2400173-aa87-476c-8216-4f8c9cf9d474-kube-api-access-scvsh\") pod \"watcher-operator-controller-manager-5db88f68c-pndq7\" (UID: \"f2400173-aa87-476c-8216-4f8c9cf9d474\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.364218 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2f9x\" (UniqueName: \"kubernetes.io/projected/0f0cbb65-3c8e-41e6-8059-34e121de0821-kube-api-access-w2f9x\") pod \"telemetry-operator-controller-manager-7f45b4ff68-5rrd9\" (UID: \"0f0cbb65-3c8e-41e6-8059-34e121de0821\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.386786 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2f9x\" (UniqueName: \"kubernetes.io/projected/0f0cbb65-3c8e-41e6-8059-34e121de0821-kube-api-access-w2f9x\") pod \"telemetry-operator-controller-manager-7f45b4ff68-5rrd9\" (UID: \"0f0cbb65-3c8e-41e6-8059-34e121de0821\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.386920 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rjvk\" (UniqueName: \"kubernetes.io/projected/0f98bc58-d0e3-405b-88fd-d8bd65f415a4-kube-api-access-8rjvk\") pod \"test-operator-controller-manager-7866795846-f7cxn\" (UID: \"0f98bc58-d0e3-405b-88fd-d8bd65f415a4\") " pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.396241 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.397285 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.420379 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.427701 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd"] Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.451307 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.470722 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scvsh\" (UniqueName: \"kubernetes.io/projected/f2400173-aa87-476c-8216-4f8c9cf9d474-kube-api-access-scvsh\") pod \"watcher-operator-controller-manager-5db88f68c-pndq7\" (UID: \"f2400173-aa87-476c-8216-4f8c9cf9d474\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.470785 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqz2k\" (UniqueName: \"kubernetes.io/projected/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-kube-api-access-pqz2k\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.470832 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.470850 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r6fn\" (UniqueName: \"kubernetes.io/projected/fe0aca33-1acf-463a-91a6-bbf35a38fd7c-kube-api-access-5r6fn\") pod \"rabbitmq-cluster-operator-manager-668c99d594-59pcs\" (UID: \"fe0aca33-1acf-463a-91a6-bbf35a38fd7c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.470891 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.470916 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:40 crc kubenswrapper[4948]: E0220 08:19:40.471097 4948 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:40 crc kubenswrapper[4948]: E0220 08:19:40.471191 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert podName:93b93501-acfe-4274-9a22-ca644b1d11d3 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:41.471165879 +0000 UTC m=+830.445660769 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert") pod "infra-operator-controller-manager-79d975b745-jrlw5" (UID: "93b93501-acfe-4274-9a22-ca644b1d11d3") : secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.490251 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scvsh\" (UniqueName: \"kubernetes.io/projected/f2400173-aa87-476c-8216-4f8c9cf9d474-kube-api-access-scvsh\") pod \"watcher-operator-controller-manager-5db88f68c-pndq7\" (UID: \"f2400173-aa87-476c-8216-4f8c9cf9d474\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.500629 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.555592 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.565675 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.572795 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqz2k\" (UniqueName: \"kubernetes.io/projected/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-kube-api-access-pqz2k\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.572881 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r6fn\" (UniqueName: \"kubernetes.io/projected/fe0aca33-1acf-463a-91a6-bbf35a38fd7c-kube-api-access-5r6fn\") pod \"rabbitmq-cluster-operator-manager-668c99d594-59pcs\" (UID: \"fe0aca33-1acf-463a-91a6-bbf35a38fd7c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.572930 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.572963 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:40 crc kubenswrapper[4948]: E0220 08:19:40.573111 4948 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 08:19:40 crc kubenswrapper[4948]: E0220 08:19:40.573176 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:41.073158109 +0000 UTC m=+830.047652929 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "webhook-server-cert" not found Feb 20 08:19:40 crc kubenswrapper[4948]: E0220 08:19:40.573230 4948 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 20 08:19:40 crc kubenswrapper[4948]: E0220 08:19:40.573255 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:41.073245821 +0000 UTC m=+830.047740651 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "metrics-server-cert" not found Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.592539 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r6fn\" (UniqueName: \"kubernetes.io/projected/fe0aca33-1acf-463a-91a6-bbf35a38fd7c-kube-api-access-5r6fn\") pod \"rabbitmq-cluster-operator-manager-668c99d594-59pcs\" (UID: \"fe0aca33-1acf-463a-91a6-bbf35a38fd7c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" Feb 20 08:19:40 crc kubenswrapper[4948]: I0220 08:19:40.594345 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqz2k\" (UniqueName: \"kubernetes.io/projected/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-kube-api-access-pqz2k\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.242494 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.255356 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.255441 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.255505 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.255615 4948 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.255687 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert podName:b8b07685-095c-4bbd-a30d-57ca59d7cbdc nodeName:}" failed. No retries permitted until 2026-02-20 08:19:42.255666653 +0000 UTC m=+831.230161473 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" (UID: "b8b07685-095c-4bbd-a30d-57ca59d7cbdc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.255802 4948 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.255866 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:42.255843947 +0000 UTC m=+831.230338837 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "webhook-server-cert" not found Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.255941 4948 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.256025 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:42.256007081 +0000 UTC m=+831.230501901 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "metrics-server-cert" not found Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.292022 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.345492 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" event={"ID":"586113f2-38b6-4bd1-8adb-3c155bb35ba9","Type":"ContainerStarted","Data":"13047a876b037c4521e3c794420c31f09e64c07bb3c2a460d3e03b1558f19e0a"} Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.366280 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-xc74s"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.386398 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.394788 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.574172 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.574389 4948 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.574454 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert podName:93b93501-acfe-4274-9a22-ca644b1d11d3 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:43.574420467 +0000 UTC m=+832.548915287 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert") pod "infra-operator-controller-manager-79d975b745-jrlw5" (UID: "93b93501-acfe-4274-9a22-ca644b1d11d3") : secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.658209 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.749478 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx"] Feb 20 08:19:41 crc kubenswrapper[4948]: W0220 08:19:41.750470 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a663826_d4b5_4ed3_8270_099b003390b4.slice/crio-d5056dc2edcf92c09e73d1ddb7aadce15676bf58b6f724db6c57780f6d5d150e WatchSource:0}: Error finding container d5056dc2edcf92c09e73d1ddb7aadce15676bf58b6f724db6c57780f6d5d150e: Status 404 returned error can't find the container with id d5056dc2edcf92c09e73d1ddb7aadce15676bf58b6f724db6c57780f6d5d150e Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.752997 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.780529 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz"] Feb 20 08:19:41 crc kubenswrapper[4948]: W0220 08:19:41.787073 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf762ec0d_e09c_4baf_9540_f4ac61ce7234.slice/crio-70b50a2a70093089e8c102ecbbaf96cb245f8d0ea9b5229b84fb487d1b462c43 WatchSource:0}: Error finding container 70b50a2a70093089e8c102ecbbaf96cb245f8d0ea9b5229b84fb487d1b462c43: Status 404 returned error can't find the container with id 70b50a2a70093089e8c102ecbbaf96cb245f8d0ea9b5229b84fb487d1b462c43 Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.803566 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.824604 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7"] Feb 20 08:19:41 crc kubenswrapper[4948]: W0220 08:19:41.827168 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod105a118c_121d_4582_960f_1da9957980cb.slice/crio-00e9662c8d49f6ee5c0d2f5b929b7f67c58779ae298ca81d10507f0502ec2bde WatchSource:0}: Error finding container 00e9662c8d49f6ee5c0d2f5b929b7f67c58779ae298ca81d10507f0502ec2bde: Status 404 returned error can't find the container with id 00e9662c8d49f6ee5c0d2f5b929b7f67c58779ae298ca81d10507f0502ec2bde Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.842581 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.855842 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.975068 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz"] Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.985668 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9"] Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.990737 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9srrq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-8497b45c89-znrkz_openstack-operators(16933747-642c-45ff-9f98-9321c633826a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 20 08:19:41 crc kubenswrapper[4948]: W0220 08:19:41.991822 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f0cbb65_3c8e_41e6_8059_34e121de0821.slice/crio-dab83244371f9d620457757e4fa587e6437cac3702cb742b7ab71532b934f994 WatchSource:0}: Error finding container dab83244371f9d620457757e4fa587e6437cac3702cb742b7ab71532b934f994: Status 404 returned error can't find the container with id dab83244371f9d620457757e4fa587e6437cac3702cb742b7ab71532b934f994 Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.991855 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" podUID="16933747-642c-45ff-9f98-9321c633826a" Feb 20 08:19:41 crc kubenswrapper[4948]: I0220 08:19:41.994762 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-5lfts"] Feb 20 08:19:41 crc kubenswrapper[4948]: W0220 08:19:41.995591 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ea675de_d1b0_4880_9652_eb066f6b0fb7.slice/crio-5652c94b03f402e891c3173cd21873fba5ae10d25a358e87372237ec7407dbdf WatchSource:0}: Error finding container 5652c94b03f402e891c3173cd21873fba5ae10d25a358e87372237ec7407dbdf: Status 404 returned error can't find the container with id 5652c94b03f402e891c3173cd21873fba5ae10d25a358e87372237ec7407dbdf Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.995634 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:66a4b9322ebb573313178ea88e31026d4532f461592b9fae2dff71efd9256d99,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w2f9x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-7f45b4ff68-5rrd9_openstack-operators(0f0cbb65-3c8e-41e6-8059-34e121de0821): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.996945 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" podUID="0f0cbb65-3c8e-41e6-8059-34e121de0821" Feb 20 08:19:41 crc kubenswrapper[4948]: W0220 08:19:41.997793 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd65e0993_11a0_4e81_963f_eeb9dcb92536.slice/crio-29ab656c5b13d0755f9f86b4e797b7aeee91cc1d92ef44299791121d7aa453db WatchSource:0}: Error finding container 29ab656c5b13d0755f9f86b4e797b7aeee91cc1d92ef44299791121d7aa453db: Status 404 returned error can't find the container with id 29ab656c5b13d0755f9f86b4e797b7aeee91cc1d92ef44299791121d7aa453db Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.998669 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q2csx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-d44cf6b75-dkkq9_openstack-operators(3ea675de-d1b0-4880-9652-eb066f6b0fb7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 20 08:19:41 crc kubenswrapper[4948]: E0220 08:19:41.999904 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" podUID="3ea675de-d1b0-4880-9652-eb066f6b0fb7" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.003401 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9"] Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.004746 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xxmkg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-68f46476f-5lfts_openstack-operators(d65e0993-11a0-4e81-963f-eeb9dcb92536): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.006110 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" podUID="d65e0993-11a0-4e81-963f-eeb9dcb92536" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.102230 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-f7cxn"] Feb 20 08:19:42 crc kubenswrapper[4948]: W0220 08:19:42.103071 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe0aca33_1acf_463a_91a6_bbf35a38fd7c.slice/crio-262d00936c737b141e11d8eb909d105b51ee27f7a50c4cdca6ed42fd7d060bdb WatchSource:0}: Error finding container 262d00936c737b141e11d8eb909d105b51ee27f7a50c4cdca6ed42fd7d060bdb: Status 404 returned error can't find the container with id 262d00936c737b141e11d8eb909d105b51ee27f7a50c4cdca6ed42fd7d060bdb Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.112294 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs"] Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.128954 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8rjvk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-7866795846-f7cxn_openstack-operators(0f98bc58-d0e3-405b-88fd-d8bd65f415a4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.130692 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" podUID="0f98bc58-d0e3-405b-88fd-d8bd65f415a4" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.133691 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7"] Feb 20 08:19:42 crc kubenswrapper[4948]: W0220 08:19:42.134965 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2400173_aa87_476c_8216_4f8c9cf9d474.slice/crio-838f9667f35a9044f976260a4ff261c3e9a746ca96302b07edae7d2a2025dcde WatchSource:0}: Error finding container 838f9667f35a9044f976260a4ff261c3e9a746ca96302b07edae7d2a2025dcde: Status 404 returned error can't find the container with id 838f9667f35a9044f976260a4ff261c3e9a746ca96302b07edae7d2a2025dcde Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.137940 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-scvsh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-5db88f68c-pndq7_openstack-operators(f2400173-aa87-476c-8216-4f8c9cf9d474): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.139277 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" podUID="f2400173-aa87-476c-8216-4f8c9cf9d474" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.286680 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.286751 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.286792 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.286822 4948 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.286880 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert podName:b8b07685-095c-4bbd-a30d-57ca59d7cbdc nodeName:}" failed. No retries permitted until 2026-02-20 08:19:44.286862282 +0000 UTC m=+833.261357102 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" (UID: "b8b07685-095c-4bbd-a30d-57ca59d7cbdc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.286921 4948 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.286943 4948 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.286999 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:44.286963744 +0000 UTC m=+833.261458614 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "metrics-server-cert" not found Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.287019 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:44.287010645 +0000 UTC m=+833.261505585 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "webhook-server-cert" not found Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.355238 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" event={"ID":"6e8e8b87-5a00-486f-a00c-1450c109f3b2","Type":"ContainerStarted","Data":"47ae6bb8643020f78c8d6b028710c10d052d0f5f885d0e6e30a064e4ee2b1808"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.359598 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" event={"ID":"347cdd7c-e5e8-49b1-a4f1-687b8a06b250","Type":"ContainerStarted","Data":"0fe0a9ff36a6df94c7f4c9e05fac924488dd011d114ed7a793ed8b30842d850b"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.360751 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" event={"ID":"3fbbd11d-612d-479a-a34f-505d995a4871","Type":"ContainerStarted","Data":"41cd793edb1558eca392ff334a39ce44122dc7272667675fbbdc34fe045cc85e"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.363496 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" event={"ID":"2d303d1a-1062-4bf6-be68-1bd6d1a3228f","Type":"ContainerStarted","Data":"58fd24496419c67ef3c86821ce7c78a64791538c2f5f99035d3770c0df3befa6"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.365055 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" event={"ID":"16933747-642c-45ff-9f98-9321c633826a","Type":"ContainerStarted","Data":"b3c67bfc9b9225d9c7af76548873d10834a152e407c56ac266aea1339f7a7815"} Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.368041 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" podUID="16933747-642c-45ff-9f98-9321c633826a" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.369319 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" event={"ID":"f762ec0d-e09c-4baf-9540-f4ac61ce7234","Type":"ContainerStarted","Data":"70b50a2a70093089e8c102ecbbaf96cb245f8d0ea9b5229b84fb487d1b462c43"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.370813 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" event={"ID":"0f98bc58-d0e3-405b-88fd-d8bd65f415a4","Type":"ContainerStarted","Data":"341ab804e809f94e8c08d8bf607b29a7ead898ee88d5194de5e3200b72af43e1"} Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.372345 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6\\\"\"" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" podUID="0f98bc58-d0e3-405b-88fd-d8bd65f415a4" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.374199 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" event={"ID":"4a663826-d4b5-4ed3-8270-099b003390b4","Type":"ContainerStarted","Data":"d5056dc2edcf92c09e73d1ddb7aadce15676bf58b6f724db6c57780f6d5d150e"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.375610 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" event={"ID":"b8e12d0c-9564-4bf2-ac61-b22d2fbdf855","Type":"ContainerStarted","Data":"6dce06766402c643cb1b6a34926fbe50490c103bdf9617d63acffc20fa7f79d4"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.393645 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" event={"ID":"fe0aca33-1acf-463a-91a6-bbf35a38fd7c","Type":"ContainerStarted","Data":"262d00936c737b141e11d8eb909d105b51ee27f7a50c4cdca6ed42fd7d060bdb"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.412907 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" event={"ID":"f2400173-aa87-476c-8216-4f8c9cf9d474","Type":"ContainerStarted","Data":"838f9667f35a9044f976260a4ff261c3e9a746ca96302b07edae7d2a2025dcde"} Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.421311 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" podUID="f2400173-aa87-476c-8216-4f8c9cf9d474" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.423590 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" event={"ID":"88142137-864d-4660-a688-a7dcc503851b","Type":"ContainerStarted","Data":"fa97191b7ad9caddbfadeaee1e4083a3f35f42d90318475a6ba08698c2cc7564"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.425776 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" event={"ID":"7c7bb531-7900-4cc6-9d9b-bae52dabc59d","Type":"ContainerStarted","Data":"69fa85ce2c62a538feaade8eead3cb5277451a5c84d034cc4d6cfdc81efb8299"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.428018 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" event={"ID":"98a1aa06-948b-4034-bc07-7e546e341a8f","Type":"ContainerStarted","Data":"e0b64b507f7813caac03a0d0569c4d3882b783f1e9f4537b9c134c856f7f94a2"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.429215 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" event={"ID":"d65e0993-11a0-4e81-963f-eeb9dcb92536","Type":"ContainerStarted","Data":"29ab656c5b13d0755f9f86b4e797b7aeee91cc1d92ef44299791121d7aa453db"} Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.432426 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" podUID="d65e0993-11a0-4e81-963f-eeb9dcb92536" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.435566 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" event={"ID":"ee33bd02-a3ce-415d-9d2d-fefd383e9810","Type":"ContainerStarted","Data":"cda502a65aa7cd9591f58265268b0fd9086950692f98e8eb2fd0370fb70d2a92"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.443014 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" event={"ID":"105a118c-121d-4582-960f-1da9957980cb","Type":"ContainerStarted","Data":"00e9662c8d49f6ee5c0d2f5b929b7f67c58779ae298ca81d10507f0502ec2bde"} Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.445011 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" event={"ID":"3ea675de-d1b0-4880-9652-eb066f6b0fb7","Type":"ContainerStarted","Data":"5652c94b03f402e891c3173cd21873fba5ae10d25a358e87372237ec7407dbdf"} Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.450467 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" podUID="3ea675de-d1b0-4880-9652-eb066f6b0fb7" Feb 20 08:19:42 crc kubenswrapper[4948]: I0220 08:19:42.450486 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" event={"ID":"0f0cbb65-3c8e-41e6-8059-34e121de0821","Type":"ContainerStarted","Data":"dab83244371f9d620457757e4fa587e6437cac3702cb742b7ab71532b934f994"} Feb 20 08:19:42 crc kubenswrapper[4948]: E0220 08:19:42.456155 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:66a4b9322ebb573313178ea88e31026d4532f461592b9fae2dff71efd9256d99\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" podUID="0f0cbb65-3c8e-41e6-8059-34e121de0821" Feb 20 08:19:43 crc kubenswrapper[4948]: E0220 08:19:43.463379 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6\\\"\"" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" podUID="0f98bc58-d0e3-405b-88fd-d8bd65f415a4" Feb 20 08:19:43 crc kubenswrapper[4948]: E0220 08:19:43.463668 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" podUID="16933747-642c-45ff-9f98-9321c633826a" Feb 20 08:19:43 crc kubenswrapper[4948]: E0220 08:19:43.463426 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" podUID="d65e0993-11a0-4e81-963f-eeb9dcb92536" Feb 20 08:19:43 crc kubenswrapper[4948]: E0220 08:19:43.463739 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:66a4b9322ebb573313178ea88e31026d4532f461592b9fae2dff71efd9256d99\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" podUID="0f0cbb65-3c8e-41e6-8059-34e121de0821" Feb 20 08:19:43 crc kubenswrapper[4948]: E0220 08:19:43.464334 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" podUID="3ea675de-d1b0-4880-9652-eb066f6b0fb7" Feb 20 08:19:43 crc kubenswrapper[4948]: E0220 08:19:43.470718 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" podUID="f2400173-aa87-476c-8216-4f8c9cf9d474" Feb 20 08:19:43 crc kubenswrapper[4948]: I0220 08:19:43.606352 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:43 crc kubenswrapper[4948]: E0220 08:19:43.606565 4948 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:43 crc kubenswrapper[4948]: E0220 08:19:43.606659 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert podName:93b93501-acfe-4274-9a22-ca644b1d11d3 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:47.606633469 +0000 UTC m=+836.581128309 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert") pod "infra-operator-controller-manager-79d975b745-jrlw5" (UID: "93b93501-acfe-4274-9a22-ca644b1d11d3") : secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:44 crc kubenswrapper[4948]: I0220 08:19:44.313961 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:44 crc kubenswrapper[4948]: I0220 08:19:44.314256 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:44 crc kubenswrapper[4948]: I0220 08:19:44.314285 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:44 crc kubenswrapper[4948]: E0220 08:19:44.314121 4948 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:44 crc kubenswrapper[4948]: E0220 08:19:44.314473 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert podName:b8b07685-095c-4bbd-a30d-57ca59d7cbdc nodeName:}" failed. No retries permitted until 2026-02-20 08:19:48.314460158 +0000 UTC m=+837.288954968 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" (UID: "b8b07685-095c-4bbd-a30d-57ca59d7cbdc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:44 crc kubenswrapper[4948]: E0220 08:19:44.314918 4948 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 20 08:19:44 crc kubenswrapper[4948]: E0220 08:19:44.314949 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:48.314940919 +0000 UTC m=+837.289435739 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "metrics-server-cert" not found Feb 20 08:19:44 crc kubenswrapper[4948]: E0220 08:19:44.314424 4948 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 08:19:44 crc kubenswrapper[4948]: E0220 08:19:44.315029 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:48.315022691 +0000 UTC m=+837.289517511 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "webhook-server-cert" not found Feb 20 08:19:47 crc kubenswrapper[4948]: I0220 08:19:47.663965 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:47 crc kubenswrapper[4948]: E0220 08:19:47.664114 4948 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:47 crc kubenswrapper[4948]: E0220 08:19:47.664195 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert podName:93b93501-acfe-4274-9a22-ca644b1d11d3 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:55.664177664 +0000 UTC m=+844.638672484 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert") pod "infra-operator-controller-manager-79d975b745-jrlw5" (UID: "93b93501-acfe-4274-9a22-ca644b1d11d3") : secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:48 crc kubenswrapper[4948]: I0220 08:19:48.375046 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:48 crc kubenswrapper[4948]: I0220 08:19:48.375546 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:48 crc kubenswrapper[4948]: E0220 08:19:48.375252 4948 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:48 crc kubenswrapper[4948]: E0220 08:19:48.375686 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert podName:b8b07685-095c-4bbd-a30d-57ca59d7cbdc nodeName:}" failed. No retries permitted until 2026-02-20 08:19:56.375661786 +0000 UTC m=+845.350156626 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" (UID: "b8b07685-095c-4bbd-a30d-57ca59d7cbdc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:48 crc kubenswrapper[4948]: I0220 08:19:48.375601 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:48 crc kubenswrapper[4948]: E0220 08:19:48.375729 4948 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 08:19:48 crc kubenswrapper[4948]: E0220 08:19:48.375744 4948 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 20 08:19:48 crc kubenswrapper[4948]: E0220 08:19:48.375813 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:56.375787459 +0000 UTC m=+845.350282359 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "webhook-server-cert" not found Feb 20 08:19:48 crc kubenswrapper[4948]: E0220 08:19:48.375839 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:19:56.37582756 +0000 UTC m=+845.350322510 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "metrics-server-cert" not found Feb 20 08:19:55 crc kubenswrapper[4948]: E0220 08:19:55.579139 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:e4689246ae78635dc3c1db9c677d8b16b8f94276df15fb9c84bfc57cc6578fcf" Feb 20 08:19:55 crc kubenswrapper[4948]: E0220 08:19:55.579757 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:e4689246ae78635dc3c1db9c677d8b16b8f94276df15fb9c84bfc57cc6578fcf,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xw4bv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-64ddbf8bb-jcnxd_openstack-operators(88142137-864d-4660-a688-a7dcc503851b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:19:55 crc kubenswrapper[4948]: E0220 08:19:55.581068 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" podUID="88142137-864d-4660-a688-a7dcc503851b" Feb 20 08:19:55 crc kubenswrapper[4948]: I0220 08:19:55.690175 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:19:55 crc kubenswrapper[4948]: E0220 08:19:55.690398 4948 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:55 crc kubenswrapper[4948]: E0220 08:19:55.690449 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert podName:93b93501-acfe-4274-9a22-ca644b1d11d3 nodeName:}" failed. No retries permitted until 2026-02-20 08:20:11.690431536 +0000 UTC m=+860.664926356 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert") pod "infra-operator-controller-manager-79d975b745-jrlw5" (UID: "93b93501-acfe-4274-9a22-ca644b1d11d3") : secret "infra-operator-webhook-server-cert" not found Feb 20 08:19:56 crc kubenswrapper[4948]: I0220 08:19:56.402401 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:19:56 crc kubenswrapper[4948]: I0220 08:19:56.402515 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:56 crc kubenswrapper[4948]: I0220 08:19:56.402614 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.402731 4948 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.402826 4948 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.402856 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:20:12.40282645 +0000 UTC m=+861.377321310 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "metrics-server-cert" not found Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.403014 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs podName:e9f13bcd-ac5a-4cbd-952a-a5b9dceee562 nodeName:}" failed. No retries permitted until 2026-02-20 08:20:12.402945743 +0000 UTC m=+861.377440603 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs") pod "openstack-operator-controller-manager-8569bc6fdb-lt4jk" (UID: "e9f13bcd-ac5a-4cbd-952a-a5b9dceee562") : secret "webhook-server-cert" not found Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.403117 4948 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.403226 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert podName:b8b07685-095c-4bbd-a30d-57ca59d7cbdc nodeName:}" failed. No retries permitted until 2026-02-20 08:20:12.403197189 +0000 UTC m=+861.377692099 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" (UID: "b8b07685-095c-4bbd-a30d-57ca59d7cbdc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.578024 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:e4689246ae78635dc3c1db9c677d8b16b8f94276df15fb9c84bfc57cc6578fcf\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" podUID="88142137-864d-4660-a688-a7dcc503851b" Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.642846 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:229fc8c8d94dd4102d2151cd4ec1eaaa09d897c2b396d06e903f61ea29c1fa34" Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.643326 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:229fc8c8d94dd4102d2151cd4ec1eaaa09d897c2b396d06e903f61ea29c1fa34,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fgpnv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-69f8888797-hl5w7_openstack-operators(ee33bd02-a3ce-415d-9d2d-fefd383e9810): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:19:56 crc kubenswrapper[4948]: E0220 08:19:56.644658 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" podUID="ee33bd02-a3ce-415d-9d2d-fefd383e9810" Feb 20 08:19:57 crc kubenswrapper[4948]: E0220 08:19:57.348789 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867" Feb 20 08:19:57 crc kubenswrapper[4948]: E0220 08:19:57.348998 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hvjkg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-554564d7fc-5fqzw_openstack-operators(105a118c-121d-4582-960f-1da9957980cb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:19:57 crc kubenswrapper[4948]: E0220 08:19:57.350260 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" podUID="105a118c-121d-4582-960f-1da9957980cb" Feb 20 08:19:57 crc kubenswrapper[4948]: E0220 08:19:57.583617 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" podUID="105a118c-121d-4582-960f-1da9957980cb" Feb 20 08:19:57 crc kubenswrapper[4948]: E0220 08:19:57.584781 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:229fc8c8d94dd4102d2151cd4ec1eaaa09d897c2b396d06e903f61ea29c1fa34\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" podUID="ee33bd02-a3ce-415d-9d2d-fefd383e9810" Feb 20 08:19:58 crc kubenswrapper[4948]: E0220 08:19:58.828394 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838" Feb 20 08:19:58 crc kubenswrapper[4948]: E0220 08:19:58.829009 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kd2fq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-567668f5cf-jxwsx_openstack-operators(7c7bb531-7900-4cc6-9d9b-bae52dabc59d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:19:58 crc kubenswrapper[4948]: E0220 08:19:58.830231 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" podUID="7c7bb531-7900-4cc6-9d9b-bae52dabc59d" Feb 20 08:19:59 crc kubenswrapper[4948]: E0220 08:19:59.226545 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Feb 20 08:19:59 crc kubenswrapper[4948]: E0220 08:19:59.226766 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5r6fn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-59pcs_openstack-operators(fe0aca33-1acf-463a-91a6-bbf35a38fd7c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:19:59 crc kubenswrapper[4948]: E0220 08:19:59.228068 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" podUID="fe0aca33-1acf-463a-91a6-bbf35a38fd7c" Feb 20 08:19:59 crc kubenswrapper[4948]: E0220 08:19:59.595841 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838\\\"\"" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" podUID="7c7bb531-7900-4cc6-9d9b-bae52dabc59d" Feb 20 08:19:59 crc kubenswrapper[4948]: E0220 08:19:59.596172 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" podUID="fe0aca33-1acf-463a-91a6-bbf35a38fd7c" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.658508 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" event={"ID":"2d303d1a-1062-4bf6-be68-1bd6d1a3228f","Type":"ContainerStarted","Data":"8b305683e46bb93e0f8a4d5f84f0b8128bce46b9e3a07adf660e42f5e58cf57a"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.659164 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.659736 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" event={"ID":"f762ec0d-e09c-4baf-9540-f4ac61ce7234","Type":"ContainerStarted","Data":"05920385456f92781604cda9d67c82eb69149c95af01964924e579fb5b7a2bad"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.659880 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.661375 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" event={"ID":"586113f2-38b6-4bd1-8adb-3c155bb35ba9","Type":"ContainerStarted","Data":"1515339e6c116f2cec374cfee4da4969c8f515a00f2e4fa528d8d741cb4b08e1"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.661528 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.663029 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" event={"ID":"f2400173-aa87-476c-8216-4f8c9cf9d474","Type":"ContainerStarted","Data":"01248ae14366e594770688ef93486b9c9108a6d178a5ad140b00869280852830"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.663232 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.664617 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" event={"ID":"98a1aa06-948b-4034-bc07-7e546e341a8f","Type":"ContainerStarted","Data":"0deb0e674afe45a81be5d52d6248d3995b7a4749e1c63e089ef0f348365dc4d0"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.664749 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.667349 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" event={"ID":"6e8e8b87-5a00-486f-a00c-1450c109f3b2","Type":"ContainerStarted","Data":"9e00325021ef73c28fdf090f3d6345fd4634b405a40c11e9d4ce76502f55fb95"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.667470 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.668712 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" event={"ID":"4a663826-d4b5-4ed3-8270-099b003390b4","Type":"ContainerStarted","Data":"4c8fb05f3bb4a17f2b52f611e5a7780744389a6989ee836a98b1d96068974f50"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.669375 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.670835 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" event={"ID":"b8e12d0c-9564-4bf2-ac61-b22d2fbdf855","Type":"ContainerStarted","Data":"4a59274e08c5951083cdb76ae3d5000b333f30161a2bc14e5de38b9e7a8ef353"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.670917 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.672374 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" event={"ID":"16933747-642c-45ff-9f98-9321c633826a","Type":"ContainerStarted","Data":"d2bcc64baae03b74eab2a38ab88906e846d18cb32b82555e43c23b89e7bb18f2"} Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.672547 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.679943 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" podStartSLOduration=10.369867812 podStartE2EDuration="29.679928382s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.406767484 +0000 UTC m=+830.381262304" lastFinishedPulling="2026-02-20 08:20:00.716828024 +0000 UTC m=+849.691322874" observedRunningTime="2026-02-20 08:20:08.676171415 +0000 UTC m=+857.650666235" watchObservedRunningTime="2026-02-20 08:20:08.679928382 +0000 UTC m=+857.654423192" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.717919 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" podStartSLOduration=10.41010731 podStartE2EDuration="29.717900667s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.407851919 +0000 UTC m=+830.382346739" lastFinishedPulling="2026-02-20 08:20:00.715645236 +0000 UTC m=+849.690140096" observedRunningTime="2026-02-20 08:20:08.712712737 +0000 UTC m=+857.687207557" watchObservedRunningTime="2026-02-20 08:20:08.717900667 +0000 UTC m=+857.692395487" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.738799 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" podStartSLOduration=10.72220704 podStartE2EDuration="29.738782328s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.700120973 +0000 UTC m=+830.674615793" lastFinishedPulling="2026-02-20 08:20:00.716696231 +0000 UTC m=+849.691191081" observedRunningTime="2026-02-20 08:20:08.735890401 +0000 UTC m=+857.710385221" watchObservedRunningTime="2026-02-20 08:20:08.738782328 +0000 UTC m=+857.713277148" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.758850 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" podStartSLOduration=3.633832566 podStartE2EDuration="29.75883161s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.990524794 +0000 UTC m=+830.965019614" lastFinishedPulling="2026-02-20 08:20:08.115523838 +0000 UTC m=+857.090018658" observedRunningTime="2026-02-20 08:20:08.758722067 +0000 UTC m=+857.733216887" watchObservedRunningTime="2026-02-20 08:20:08.75883161 +0000 UTC m=+857.733326430" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.794647 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" podStartSLOduration=10.505604491 podStartE2EDuration="29.794624975s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.427789909 +0000 UTC m=+830.402284729" lastFinishedPulling="2026-02-20 08:20:00.716810393 +0000 UTC m=+849.691305213" observedRunningTime="2026-02-20 08:20:08.783685343 +0000 UTC m=+857.758180163" watchObservedRunningTime="2026-02-20 08:20:08.794624975 +0000 UTC m=+857.769119795" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.840499 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" podStartSLOduration=10.878320488 podStartE2EDuration="29.840479741s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.754560058 +0000 UTC m=+830.729054878" lastFinishedPulling="2026-02-20 08:20:00.716719271 +0000 UTC m=+849.691214131" observedRunningTime="2026-02-20 08:20:08.815087936 +0000 UTC m=+857.789582756" watchObservedRunningTime="2026-02-20 08:20:08.840479741 +0000 UTC m=+857.814974561" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.841555 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" podStartSLOduration=10.916095548 podStartE2EDuration="29.841550506s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.79113891 +0000 UTC m=+830.765633730" lastFinishedPulling="2026-02-20 08:20:00.716593858 +0000 UTC m=+849.691088688" observedRunningTime="2026-02-20 08:20:08.840666055 +0000 UTC m=+857.815160875" watchObservedRunningTime="2026-02-20 08:20:08.841550506 +0000 UTC m=+857.816045326" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.868165 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" podStartSLOduration=2.890135941 podStartE2EDuration="28.868141718s" podCreationTimestamp="2026-02-20 08:19:40 +0000 UTC" firstStartedPulling="2026-02-20 08:19:42.137636894 +0000 UTC m=+831.112131714" lastFinishedPulling="2026-02-20 08:20:08.115642641 +0000 UTC m=+857.090137491" observedRunningTime="2026-02-20 08:20:08.865083558 +0000 UTC m=+857.839578378" watchObservedRunningTime="2026-02-20 08:20:08.868141718 +0000 UTC m=+857.842636538" Feb 20 08:20:08 crc kubenswrapper[4948]: I0220 08:20:08.940939 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" podStartSLOduration=11.193230774 podStartE2EDuration="29.940920625s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:40.457184817 +0000 UTC m=+829.431679627" lastFinishedPulling="2026-02-20 08:19:59.204874618 +0000 UTC m=+848.179369478" observedRunningTime="2026-02-20 08:20:08.897118816 +0000 UTC m=+857.871613646" watchObservedRunningTime="2026-02-20 08:20:08.940920625 +0000 UTC m=+857.915415445" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.679870 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" event={"ID":"105a118c-121d-4582-960f-1da9957980cb","Type":"ContainerStarted","Data":"a3e9593c44b650a8bda80c835d18ab928cf541897ad72ff38a186676ab68f51a"} Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.680653 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.681623 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" event={"ID":"347cdd7c-e5e8-49b1-a4f1-687b8a06b250","Type":"ContainerStarted","Data":"39750772c75580cafb02d68765f21c5eb0ff305d0b5a17c3eb5102bfe7d9adaa"} Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.681937 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.682965 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" event={"ID":"3ea675de-d1b0-4880-9652-eb066f6b0fb7","Type":"ContainerStarted","Data":"6eb878f02b7a4138e137c525a1a7da5d56bf0e67f0c5146cbf73c71230b4ec14"} Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.683148 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.684296 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" event={"ID":"3fbbd11d-612d-479a-a34f-505d995a4871","Type":"ContainerStarted","Data":"915dd46835ad50adb4f7eb94e42e8320fcf29cc6712abcb40c93bf0df022e529"} Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.684613 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.686154 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" event={"ID":"0f0cbb65-3c8e-41e6-8059-34e121de0821","Type":"ContainerStarted","Data":"87debc5d746918ca7f6810ce45f0ea2507d75c0c6d9365af94ce2ebc25747200"} Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.686489 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.687445 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" event={"ID":"0f98bc58-d0e3-405b-88fd-d8bd65f415a4","Type":"ContainerStarted","Data":"e8006548cce48d3fe0f5cc75d22075761038af6d79ccedbe3d91f697f3734ee2"} Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.687755 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.688714 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" event={"ID":"d65e0993-11a0-4e81-963f-eeb9dcb92536","Type":"ContainerStarted","Data":"d630d43c3b24ea13aaf65ad43bba3892eb33c2f90cc3763b8fb4dbc270ae7add"} Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.689028 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.690201 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" event={"ID":"ee33bd02-a3ce-415d-9d2d-fefd383e9810","Type":"ContainerStarted","Data":"89e896c05dd7dac8b33eb14ad96e8b0e495eb55fb5db898aab1550650fce14f3"} Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.690492 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.706751 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" podStartSLOduration=3.44130413 podStartE2EDuration="30.706736469s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.844857818 +0000 UTC m=+830.819352638" lastFinishedPulling="2026-02-20 08:20:09.110290157 +0000 UTC m=+858.084784977" observedRunningTime="2026-02-20 08:20:09.70288596 +0000 UTC m=+858.677380780" watchObservedRunningTime="2026-02-20 08:20:09.706736469 +0000 UTC m=+858.681231289" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.741426 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" podStartSLOduration=4.497568096 podStartE2EDuration="30.741410208s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:42.004640559 +0000 UTC m=+830.979135379" lastFinishedPulling="2026-02-20 08:20:08.248482651 +0000 UTC m=+857.222977491" observedRunningTime="2026-02-20 08:20:09.736601647 +0000 UTC m=+858.711096457" watchObservedRunningTime="2026-02-20 08:20:09.741410208 +0000 UTC m=+858.715905028" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.758565 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" podStartSLOduration=3.359207277 podStartE2EDuration="30.758549572s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.813986867 +0000 UTC m=+830.788481687" lastFinishedPulling="2026-02-20 08:20:09.213329162 +0000 UTC m=+858.187823982" observedRunningTime="2026-02-20 08:20:09.758071471 +0000 UTC m=+858.732566291" watchObservedRunningTime="2026-02-20 08:20:09.758549572 +0000 UTC m=+858.733044412" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.788526 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" podStartSLOduration=3.800824093 podStartE2EDuration="29.788508903s" podCreationTimestamp="2026-02-20 08:19:40 +0000 UTC" firstStartedPulling="2026-02-20 08:19:42.128767859 +0000 UTC m=+831.103262679" lastFinishedPulling="2026-02-20 08:20:08.116452669 +0000 UTC m=+857.090947489" observedRunningTime="2026-02-20 08:20:09.782783801 +0000 UTC m=+858.757278621" watchObservedRunningTime="2026-02-20 08:20:09.788508903 +0000 UTC m=+858.763003723" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.831568 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" podStartSLOduration=13.03607257 podStartE2EDuration="30.831549014s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.409542988 +0000 UTC m=+830.384037808" lastFinishedPulling="2026-02-20 08:19:59.205019432 +0000 UTC m=+848.179514252" observedRunningTime="2026-02-20 08:20:09.830281615 +0000 UTC m=+858.804776425" watchObservedRunningTime="2026-02-20 08:20:09.831549014 +0000 UTC m=+858.806043834" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.856002 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" podStartSLOduration=11.991700769 podStartE2EDuration="30.855984537s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.852362551 +0000 UTC m=+830.826857361" lastFinishedPulling="2026-02-20 08:20:00.716646309 +0000 UTC m=+849.691141129" observedRunningTime="2026-02-20 08:20:09.850562442 +0000 UTC m=+858.825057262" watchObservedRunningTime="2026-02-20 08:20:09.855984537 +0000 UTC m=+858.830479357" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.894961 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" podStartSLOduration=4.596365751 podStartE2EDuration="30.894947345s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.995486828 +0000 UTC m=+830.969981638" lastFinishedPulling="2026-02-20 08:20:08.294068402 +0000 UTC m=+857.268563232" observedRunningTime="2026-02-20 08:20:09.892523219 +0000 UTC m=+858.867018039" watchObservedRunningTime="2026-02-20 08:20:09.894947345 +0000 UTC m=+858.869442165" Feb 20 08:20:09 crc kubenswrapper[4948]: I0220 08:20:09.923101 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" podStartSLOduration=4.661345039 podStartE2EDuration="30.923084393s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.99858411 +0000 UTC m=+830.973078920" lastFinishedPulling="2026-02-20 08:20:08.260323444 +0000 UTC m=+857.234818274" observedRunningTime="2026-02-20 08:20:09.920024893 +0000 UTC m=+858.894519713" watchObservedRunningTime="2026-02-20 08:20:09.923084393 +0000 UTC m=+858.897579213" Feb 20 08:20:11 crc kubenswrapper[4948]: I0220 08:20:11.705103 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" event={"ID":"7c7bb531-7900-4cc6-9d9b-bae52dabc59d","Type":"ContainerStarted","Data":"6bf4067c3cc3f27ae6ff2e15a723b9f8e806372c7cfdf33156af0d24a9c1de0b"} Feb 20 08:20:11 crc kubenswrapper[4948]: I0220 08:20:11.705696 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" Feb 20 08:20:11 crc kubenswrapper[4948]: I0220 08:20:11.729620 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" podStartSLOduration=3.321213943 podStartE2EDuration="32.729581945s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.750419332 +0000 UTC m=+830.724914152" lastFinishedPulling="2026-02-20 08:20:11.158787334 +0000 UTC m=+860.133282154" observedRunningTime="2026-02-20 08:20:11.72241699 +0000 UTC m=+860.696911810" watchObservedRunningTime="2026-02-20 08:20:11.729581945 +0000 UTC m=+860.704076805" Feb 20 08:20:11 crc kubenswrapper[4948]: I0220 08:20:11.744010 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:20:11 crc kubenswrapper[4948]: I0220 08:20:11.758798 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93b93501-acfe-4274-9a22-ca644b1d11d3-cert\") pod \"infra-operator-controller-manager-79d975b745-jrlw5\" (UID: \"93b93501-acfe-4274-9a22-ca644b1d11d3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:20:11 crc kubenswrapper[4948]: I0220 08:20:11.818122 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-nxfl4" Feb 20 08:20:11 crc kubenswrapper[4948]: I0220 08:20:11.825884 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.247815 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5"] Feb 20 08:20:12 crc kubenswrapper[4948]: W0220 08:20:12.266743 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93b93501_acfe_4274_9a22_ca644b1d11d3.slice/crio-56ecd22253c889793ab9c7e5698f95c754daf76da3e78e23c40c4217acbf5737 WatchSource:0}: Error finding container 56ecd22253c889793ab9c7e5698f95c754daf76da3e78e23c40c4217acbf5737: Status 404 returned error can't find the container with id 56ecd22253c889793ab9c7e5698f95c754daf76da3e78e23c40c4217acbf5737 Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.454351 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.454497 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.454585 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.467444 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-webhook-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.467451 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9f13bcd-ac5a-4cbd-952a-a5b9dceee562-metrics-certs\") pod \"openstack-operator-controller-manager-8569bc6fdb-lt4jk\" (UID: \"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562\") " pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.469061 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b8b07685-095c-4bbd-a30d-57ca59d7cbdc-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m\" (UID: \"b8b07685-095c-4bbd-a30d-57ca59d7cbdc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.709568 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-5sbbv" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.713886 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" event={"ID":"93b93501-acfe-4274-9a22-ca644b1d11d3","Type":"ContainerStarted","Data":"56ecd22253c889793ab9c7e5698f95c754daf76da3e78e23c40c4217acbf5737"} Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.716717 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" event={"ID":"88142137-864d-4660-a688-a7dcc503851b","Type":"ContainerStarted","Data":"c6f25d4bd83aaf20517efd9ccdb7c5ba8f2abb24d30560a0ea4e0eba2e56aec0"} Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.717668 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.717787 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.746649 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" podStartSLOduration=3.423646154 podStartE2EDuration="33.746618318s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:19:41.81413517 +0000 UTC m=+830.788629990" lastFinishedPulling="2026-02-20 08:20:12.137107334 +0000 UTC m=+861.111602154" observedRunningTime="2026-02-20 08:20:12.74280354 +0000 UTC m=+861.717298360" watchObservedRunningTime="2026-02-20 08:20:12.746618318 +0000 UTC m=+861.721113178" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.767932 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-c7nf5" Feb 20 08:20:12 crc kubenswrapper[4948]: I0220 08:20:12.776936 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.000309 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk"] Feb 20 08:20:13 crc kubenswrapper[4948]: W0220 08:20:13.013687 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9f13bcd_ac5a_4cbd_952a_a5b9dceee562.slice/crio-80bd3ea908fb4eb09f8f4416da6a368cb8331471b5e64de7116e12d5a966ea56 WatchSource:0}: Error finding container 80bd3ea908fb4eb09f8f4416da6a368cb8331471b5e64de7116e12d5a966ea56: Status 404 returned error can't find the container with id 80bd3ea908fb4eb09f8f4416da6a368cb8331471b5e64de7116e12d5a966ea56 Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.284209 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m"] Feb 20 08:20:13 crc kubenswrapper[4948]: W0220 08:20:13.300266 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8b07685_095c_4bbd_a30d_57ca59d7cbdc.slice/crio-95d33a746d766d859f961c9d4ead205a2fa3e2416d7051db3d97c9a00f159cfa WatchSource:0}: Error finding container 95d33a746d766d859f961c9d4ead205a2fa3e2416d7051db3d97c9a00f159cfa: Status 404 returned error can't find the container with id 95d33a746d766d859f961c9d4ead205a2fa3e2416d7051db3d97c9a00f159cfa Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.739346 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" event={"ID":"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562","Type":"ContainerStarted","Data":"0a26ac33ec17bf06b70ad74d59f7ee41394a2704c8d4b99685e4a4ce9ccf41fa"} Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.739384 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" event={"ID":"e9f13bcd-ac5a-4cbd-952a-a5b9dceee562","Type":"ContainerStarted","Data":"80bd3ea908fb4eb09f8f4416da6a368cb8331471b5e64de7116e12d5a966ea56"} Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.739397 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" event={"ID":"b8b07685-095c-4bbd-a30d-57ca59d7cbdc","Type":"ContainerStarted","Data":"95d33a746d766d859f961c9d4ead205a2fa3e2416d7051db3d97c9a00f159cfa"} Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.739410 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" event={"ID":"fe0aca33-1acf-463a-91a6-bbf35a38fd7c","Type":"ContainerStarted","Data":"b5e51c8afb7e870e869321f303b239d0d185612855b1179878e8aaf17edaffcb"} Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.739426 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.755931 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" podStartSLOduration=33.755916171 podStartE2EDuration="33.755916171s" podCreationTimestamp="2026-02-20 08:19:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:20:13.75459106 +0000 UTC m=+862.729085910" watchObservedRunningTime="2026-02-20 08:20:13.755916171 +0000 UTC m=+862.730410991" Feb 20 08:20:13 crc kubenswrapper[4948]: I0220 08:20:13.784409 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-59pcs" podStartSLOduration=2.582149195 podStartE2EDuration="33.784391917s" podCreationTimestamp="2026-02-20 08:19:40 +0000 UTC" firstStartedPulling="2026-02-20 08:19:42.106879305 +0000 UTC m=+831.081374125" lastFinishedPulling="2026-02-20 08:20:13.309122007 +0000 UTC m=+862.283616847" observedRunningTime="2026-02-20 08:20:13.777319044 +0000 UTC m=+862.751813874" watchObservedRunningTime="2026-02-20 08:20:13.784391917 +0000 UTC m=+862.758886747" Feb 20 08:20:14 crc kubenswrapper[4948]: I0220 08:20:14.744114 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" event={"ID":"93b93501-acfe-4274-9a22-ca644b1d11d3","Type":"ContainerStarted","Data":"118b2eef7d5016594db3982e89ab763d690309404c4d622207ee663120afdae1"} Feb 20 08:20:14 crc kubenswrapper[4948]: I0220 08:20:14.768250 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" podStartSLOduration=33.66335475 podStartE2EDuration="35.768229535s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:20:12.26928645 +0000 UTC m=+861.243781270" lastFinishedPulling="2026-02-20 08:20:14.374161235 +0000 UTC m=+863.348656055" observedRunningTime="2026-02-20 08:20:14.757200981 +0000 UTC m=+863.731695811" watchObservedRunningTime="2026-02-20 08:20:14.768229535 +0000 UTC m=+863.742724355" Feb 20 08:20:15 crc kubenswrapper[4948]: I0220 08:20:15.751849 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:20:16 crc kubenswrapper[4948]: I0220 08:20:16.761573 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" event={"ID":"b8b07685-095c-4bbd-a30d-57ca59d7cbdc","Type":"ContainerStarted","Data":"3e13e2e5df6fe5e1ffd0e5435f05a87967318a31abfc6a13ad7a56936b7cec37"} Feb 20 08:20:16 crc kubenswrapper[4948]: I0220 08:20:16.762061 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:20:16 crc kubenswrapper[4948]: I0220 08:20:16.804920 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" podStartSLOduration=35.346814965 podStartE2EDuration="37.804897129s" podCreationTimestamp="2026-02-20 08:19:39 +0000 UTC" firstStartedPulling="2026-02-20 08:20:13.309294291 +0000 UTC m=+862.283789111" lastFinishedPulling="2026-02-20 08:20:15.767376445 +0000 UTC m=+864.741871275" observedRunningTime="2026-02-20 08:20:16.796966206 +0000 UTC m=+865.771461076" watchObservedRunningTime="2026-02-20 08:20:16.804897129 +0000 UTC m=+865.779391949" Feb 20 08:20:19 crc kubenswrapper[4948]: I0220 08:20:19.873269 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-4h9v5" Feb 20 08:20:19 crc kubenswrapper[4948]: I0220 08:20:19.886738 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c6xqd" Feb 20 08:20:19 crc kubenswrapper[4948]: I0220 08:20:19.907781 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-kjpkr" Feb 20 08:20:19 crc kubenswrapper[4948]: I0220 08:20:19.974460 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-6cr5p" Feb 20 08:20:19 crc kubenswrapper[4948]: I0220 08:20:19.986887 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987464f4-xc74s" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.035258 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-bkl49" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.058379 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-5fqzw" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.090740 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-lj2sz" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.158634 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-ttkfm" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.181225 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-zcccd" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.263132 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-jcnxd" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.286947 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-jxwsx" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.323616 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-hl5w7" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.400287 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-dkkq9" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.422822 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-znrkz" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.454124 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5lfts" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.503887 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-5rrd9" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.558776 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-7866795846-f7cxn" Feb 20 08:20:20 crc kubenswrapper[4948]: I0220 08:20:20.572848 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-pndq7" Feb 20 08:20:21 crc kubenswrapper[4948]: I0220 08:20:21.836534 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-79d975b745-jrlw5" Feb 20 08:20:22 crc kubenswrapper[4948]: I0220 08:20:22.726030 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-8569bc6fdb-lt4jk" Feb 20 08:20:22 crc kubenswrapper[4948]: I0220 08:20:22.794334 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m" Feb 20 08:20:38 crc kubenswrapper[4948]: I0220 08:20:38.025017 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:20:38 crc kubenswrapper[4948]: I0220 08:20:38.025472 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.757659 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9dfhs"] Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.759022 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.763467 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.763682 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.763827 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.764033 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-qgm2g" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.777877 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9dfhs"] Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.815524 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cqgz2"] Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.816561 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.819804 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.835203 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cqgz2"] Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.890404 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d754\" (UniqueName: \"kubernetes.io/projected/1224cf30-813c-48f2-92c9-fabe78bbc6bc-kube-api-access-7d754\") pod \"dnsmasq-dns-675f4bcbfc-9dfhs\" (UID: \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.890599 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1224cf30-813c-48f2-92c9-fabe78bbc6bc-config\") pod \"dnsmasq-dns-675f4bcbfc-9dfhs\" (UID: \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.991552 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d754\" (UniqueName: \"kubernetes.io/projected/1224cf30-813c-48f2-92c9-fabe78bbc6bc-kube-api-access-7d754\") pod \"dnsmasq-dns-675f4bcbfc-9dfhs\" (UID: \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.991632 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffpsp\" (UniqueName: \"kubernetes.io/projected/05b1d586-95db-4795-a032-f22ba90e6c16-kube-api-access-ffpsp\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.991680 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.991703 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-config\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.991743 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1224cf30-813c-48f2-92c9-fabe78bbc6bc-config\") pod \"dnsmasq-dns-675f4bcbfc-9dfhs\" (UID: \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:20:39 crc kubenswrapper[4948]: I0220 08:20:39.992517 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1224cf30-813c-48f2-92c9-fabe78bbc6bc-config\") pod \"dnsmasq-dns-675f4bcbfc-9dfhs\" (UID: \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.009194 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d754\" (UniqueName: \"kubernetes.io/projected/1224cf30-813c-48f2-92c9-fabe78bbc6bc-kube-api-access-7d754\") pod \"dnsmasq-dns-675f4bcbfc-9dfhs\" (UID: \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.080862 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.093360 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffpsp\" (UniqueName: \"kubernetes.io/projected/05b1d586-95db-4795-a032-f22ba90e6c16-kube-api-access-ffpsp\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.093451 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.093489 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-config\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.094566 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.094565 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-config\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.125916 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffpsp\" (UniqueName: \"kubernetes.io/projected/05b1d586-95db-4795-a032-f22ba90e6c16-kube-api-access-ffpsp\") pod \"dnsmasq-dns-78dd6ddcc-cqgz2\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.133963 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.598701 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9dfhs"] Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.599947 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:20:40 crc kubenswrapper[4948]: W0220 08:20:40.608279 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05b1d586_95db_4795_a032_f22ba90e6c16.slice/crio-4e5e167063afcb1f5336b93cee715fd831f959a4a5fb12e9e562274ec18ad672 WatchSource:0}: Error finding container 4e5e167063afcb1f5336b93cee715fd831f959a4a5fb12e9e562274ec18ad672: Status 404 returned error can't find the container with id 4e5e167063afcb1f5336b93cee715fd831f959a4a5fb12e9e562274ec18ad672 Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.612651 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cqgz2"] Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.969921 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" event={"ID":"1224cf30-813c-48f2-92c9-fabe78bbc6bc","Type":"ContainerStarted","Data":"459736be569ebf9acfc7e517d56782883ab07764b913b8c4d305735259e95b55"} Feb 20 08:20:40 crc kubenswrapper[4948]: I0220 08:20:40.971360 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" event={"ID":"05b1d586-95db-4795-a032-f22ba90e6c16","Type":"ContainerStarted","Data":"4e5e167063afcb1f5336b93cee715fd831f959a4a5fb12e9e562274ec18ad672"} Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.526616 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9dfhs"] Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.564921 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zt6gg"] Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.566566 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.572546 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zt6gg"] Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.642457 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pds7s\" (UniqueName: \"kubernetes.io/projected/977992e9-30ca-4420-98ae-92f91545cd10-kube-api-access-pds7s\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.642941 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.643076 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-config\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.743863 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pds7s\" (UniqueName: \"kubernetes.io/projected/977992e9-30ca-4420-98ae-92f91545cd10-kube-api-access-pds7s\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.743925 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.743989 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-config\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.745146 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-config\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.746071 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.765593 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pds7s\" (UniqueName: \"kubernetes.io/projected/977992e9-30ca-4420-98ae-92f91545cd10-kube-api-access-pds7s\") pod \"dnsmasq-dns-666b6646f7-zt6gg\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.879069 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cqgz2"] Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.895850 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.921143 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mkkzf"] Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.927271 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:42 crc kubenswrapper[4948]: I0220 08:20:42.941611 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mkkzf"] Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.048022 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdnm9\" (UniqueName: \"kubernetes.io/projected/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-kube-api-access-hdnm9\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.048096 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-config\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.048146 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.149790 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-config\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.149864 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.149914 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdnm9\" (UniqueName: \"kubernetes.io/projected/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-kube-api-access-hdnm9\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.150868 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-config\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.151096 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.173997 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdnm9\" (UniqueName: \"kubernetes.io/projected/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-kube-api-access-hdnm9\") pod \"dnsmasq-dns-57d769cc4f-mkkzf\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.316471 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.424430 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zt6gg"] Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.774450 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.776295 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.778663 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.783628 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.783833 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.783951 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.784072 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-8zcmc" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.784175 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.784234 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.797520 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.867695 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mkkzf"] Feb 20 08:20:43 crc kubenswrapper[4948]: W0220 08:20:43.882387 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5686ae39_4e44_4303_9b44_ad4b9ceae1ba.slice/crio-b4980c083e552022d3673e8c44cd0e8286ee33627cbef5030e75bd7be40133d3 WatchSource:0}: Error finding container b4980c083e552022d3673e8c44cd0e8286ee33627cbef5030e75bd7be40133d3: Status 404 returned error can't find the container with id b4980c083e552022d3673e8c44cd0e8286ee33627cbef5030e75bd7be40133d3 Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.963412 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/73b84bb7-f594-4823-ac03-40fdac6ee177-pod-info\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.963520 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqlcq\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-kube-api-access-tqlcq\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.963663 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.963709 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.963825 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/73b84bb7-f594-4823-ac03-40fdac6ee177-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.964835 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-server-conf\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.964918 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.965133 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-config-data\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.965228 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.965330 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:43 crc kubenswrapper[4948]: I0220 08:20:43.965347 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.006642 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" event={"ID":"5686ae39-4e44-4303-9b44-ad4b9ceae1ba","Type":"ContainerStarted","Data":"b4980c083e552022d3673e8c44cd0e8286ee33627cbef5030e75bd7be40133d3"} Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.008015 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" event={"ID":"977992e9-30ca-4420-98ae-92f91545cd10","Type":"ContainerStarted","Data":"a0508d6a1f4b7f39846aac7f90bdfe83ad8793d4225f8bb5f2eaca0710969807"} Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.068987 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069047 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069063 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069092 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/73b84bb7-f594-4823-ac03-40fdac6ee177-pod-info\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069123 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqlcq\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-kube-api-access-tqlcq\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069139 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069162 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069178 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/73b84bb7-f594-4823-ac03-40fdac6ee177-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069200 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-server-conf\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069226 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069251 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-config-data\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069477 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.069910 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.070699 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.070748 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-config-data\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.070880 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.071231 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-server-conf\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.079019 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/73b84bb7-f594-4823-ac03-40fdac6ee177-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.080674 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/73b84bb7-f594-4823-ac03-40fdac6ee177-pod-info\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.082032 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.084316 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.100691 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqlcq\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-kube-api-access-tqlcq\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.107661 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.116465 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.132486 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.145886 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.146007 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.149255 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.149449 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.149560 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-rrjt5" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.149663 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.149758 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.149873 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.150067 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203624 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203708 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203749 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203790 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203812 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203845 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203879 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7t2f\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-kube-api-access-b7t2f\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203904 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203946 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/790894c9-053c-497c-955d-ce7519111dd6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.203964 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.204889 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/790894c9-053c-497c-955d-ce7519111dd6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.309839 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.309901 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.309941 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.309998 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7t2f\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-kube-api-access-b7t2f\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.310030 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.310058 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/790894c9-053c-497c-955d-ce7519111dd6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.310087 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.310111 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/790894c9-053c-497c-955d-ce7519111dd6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.310147 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.310207 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.310247 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.310638 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.311291 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.311865 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.312108 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.313112 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.313149 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.313399 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.317290 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/790894c9-053c-497c-955d-ce7519111dd6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.317461 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/790894c9-053c-497c-955d-ce7519111dd6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.330722 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.331723 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.331845 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7t2f\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-kube-api-access-b7t2f\") pod \"rabbitmq-cell1-server-0\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.397123 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:20:44 crc kubenswrapper[4948]: W0220 08:20:44.403661 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73b84bb7_f594_4823_ac03_40fdac6ee177.slice/crio-e552b8693d7f82c1ed5a2c7c103ec119814c444b37d5f93e2491692378454a4c WatchSource:0}: Error finding container e552b8693d7f82c1ed5a2c7c103ec119814c444b37d5f93e2491692378454a4c: Status 404 returned error can't find the container with id e552b8693d7f82c1ed5a2c7c103ec119814c444b37d5f93e2491692378454a4c Feb 20 08:20:44 crc kubenswrapper[4948]: I0220 08:20:44.496737 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.027931 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"73b84bb7-f594-4823-ac03-40fdac6ee177","Type":"ContainerStarted","Data":"e552b8693d7f82c1ed5a2c7c103ec119814c444b37d5f93e2491692378454a4c"} Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.186081 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.191312 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.194287 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.194407 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-w44rf" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.194534 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.194868 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.200065 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.202025 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.325407 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-kolla-config\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.325503 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6e59b63d-0c56-488c-87cb-348af87058c6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.325524 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e59b63d-0c56-488c-87cb-348af87058c6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.325564 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e59b63d-0c56-488c-87cb-348af87058c6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.325593 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq68c\" (UniqueName: \"kubernetes.io/projected/6e59b63d-0c56-488c-87cb-348af87058c6-kube-api-access-zq68c\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.325610 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.325635 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-config-data-default\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.325654 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.431688 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-kolla-config\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.431963 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6e59b63d-0c56-488c-87cb-348af87058c6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.431996 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e59b63d-0c56-488c-87cb-348af87058c6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.432042 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e59b63d-0c56-488c-87cb-348af87058c6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.432077 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq68c\" (UniqueName: \"kubernetes.io/projected/6e59b63d-0c56-488c-87cb-348af87058c6-kube-api-access-zq68c\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.432114 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.432147 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-config-data-default\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.432165 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.432348 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rt8tp"] Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.434047 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.436565 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.437301 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.437537 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-config-data-default\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.437542 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6e59b63d-0c56-488c-87cb-348af87058c6-kolla-config\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.440215 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6e59b63d-0c56-488c-87cb-348af87058c6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.464657 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rt8tp"] Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.481852 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e59b63d-0c56-488c-87cb-348af87058c6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.483259 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq68c\" (UniqueName: \"kubernetes.io/projected/6e59b63d-0c56-488c-87cb-348af87058c6-kube-api-access-zq68c\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.492028 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.498276 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e59b63d-0c56-488c-87cb-348af87058c6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6e59b63d-0c56-488c-87cb-348af87058c6\") " pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.518315 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.635050 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-utilities\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.636641 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2cfc\" (UniqueName: \"kubernetes.io/projected/93f24735-d27d-4867-b9f7-cce47e7f376f-kube-api-access-q2cfc\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.636704 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-catalog-content\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.738346 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-utilities\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.738400 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2cfc\" (UniqueName: \"kubernetes.io/projected/93f24735-d27d-4867-b9f7-cce47e7f376f-kube-api-access-q2cfc\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.738444 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-catalog-content\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.739092 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-utilities\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.739126 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-catalog-content\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.757178 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2cfc\" (UniqueName: \"kubernetes.io/projected/93f24735-d27d-4867-b9f7-cce47e7f376f-kube-api-access-q2cfc\") pod \"redhat-operators-rt8tp\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:45 crc kubenswrapper[4948]: I0220 08:20:45.852465 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.562889 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.564163 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.574291 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-wmf44" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.574688 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.575040 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.575178 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.582064 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.654992 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qwvn\" (UniqueName: \"kubernetes.io/projected/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-kube-api-access-7qwvn\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.655069 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.655102 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.655124 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.655139 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.655159 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.655237 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.655273 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.756697 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.756990 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.757031 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qwvn\" (UniqueName: \"kubernetes.io/projected/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-kube-api-access-7qwvn\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.757067 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.757097 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.757119 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.757135 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.757157 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.757834 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.758071 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.760199 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.760541 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.763526 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.763769 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.768983 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.769865 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.778018 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.780220 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.780401 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.780506 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-4mvzk" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.781431 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.795103 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.805439 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qwvn\" (UniqueName: \"kubernetes.io/projected/d02114ae-21f2-4cc5-9bca-9505c25eaf5f-kube-api-access-7qwvn\") pod \"openstack-cell1-galera-0\" (UID: \"d02114ae-21f2-4cc5-9bca-9505c25eaf5f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.858522 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-config-data\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.858602 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mngjx\" (UniqueName: \"kubernetes.io/projected/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-kube-api-access-mngjx\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.858629 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.858647 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.858673 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-kolla-config\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.897168 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.959479 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-config-data\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.959544 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mngjx\" (UniqueName: \"kubernetes.io/projected/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-kube-api-access-mngjx\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.959574 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.959592 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.959610 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-kolla-config\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.960248 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-kolla-config\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.960586 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-config-data\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.962991 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.966596 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:46 crc kubenswrapper[4948]: I0220 08:20:46.976118 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mngjx\" (UniqueName: \"kubernetes.io/projected/8e994489-a2cf-4f6d-a00c-98f627ba0e5f-kube-api-access-mngjx\") pod \"memcached-0\" (UID: \"8e994489-a2cf-4f6d-a00c-98f627ba0e5f\") " pod="openstack/memcached-0" Feb 20 08:20:47 crc kubenswrapper[4948]: I0220 08:20:47.147203 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 20 08:20:49 crc kubenswrapper[4948]: I0220 08:20:49.040986 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:20:49 crc kubenswrapper[4948]: I0220 08:20:49.042380 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 08:20:49 crc kubenswrapper[4948]: I0220 08:20:49.044748 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-d57tg" Feb 20 08:20:49 crc kubenswrapper[4948]: I0220 08:20:49.052341 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:20:49 crc kubenswrapper[4948]: I0220 08:20:49.204133 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqgmb\" (UniqueName: \"kubernetes.io/projected/0dfbb200-e444-4895-a410-e51cbb1fc112-kube-api-access-rqgmb\") pod \"kube-state-metrics-0\" (UID: \"0dfbb200-e444-4895-a410-e51cbb1fc112\") " pod="openstack/kube-state-metrics-0" Feb 20 08:20:49 crc kubenswrapper[4948]: I0220 08:20:49.305693 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqgmb\" (UniqueName: \"kubernetes.io/projected/0dfbb200-e444-4895-a410-e51cbb1fc112-kube-api-access-rqgmb\") pod \"kube-state-metrics-0\" (UID: \"0dfbb200-e444-4895-a410-e51cbb1fc112\") " pod="openstack/kube-state-metrics-0" Feb 20 08:20:49 crc kubenswrapper[4948]: I0220 08:20:49.326673 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqgmb\" (UniqueName: \"kubernetes.io/projected/0dfbb200-e444-4895-a410-e51cbb1fc112-kube-api-access-rqgmb\") pod \"kube-state-metrics-0\" (UID: \"0dfbb200-e444-4895-a410-e51cbb1fc112\") " pod="openstack/kube-state-metrics-0" Feb 20 08:20:49 crc kubenswrapper[4948]: I0220 08:20:49.373628 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.844697 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-mkpng"] Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.845756 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mkpng" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.851278 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-258bh" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.851595 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.854076 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.854081 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-q4bv4"] Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.856694 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.873295 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-q4bv4"] Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.890033 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mkpng"] Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.953901 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-log\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.953965 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b80cb988-de2a-4e65-9161-a0af0561c754-ovn-controller-tls-certs\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954012 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-run\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954088 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80cb988-de2a-4e65-9161-a0af0561c754-combined-ca-bundle\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954251 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-run\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954326 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-log-ovn\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954356 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-run-ovn\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954383 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c82772b6-9737-45ee-998a-e897086c03b5-scripts\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954424 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b80cb988-de2a-4e65-9161-a0af0561c754-scripts\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954461 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn4rx\" (UniqueName: \"kubernetes.io/projected/c82772b6-9737-45ee-998a-e897086c03b5-kube-api-access-hn4rx\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954524 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-etc-ovs\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954657 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5ql2\" (UniqueName: \"kubernetes.io/projected/b80cb988-de2a-4e65-9161-a0af0561c754-kube-api-access-h5ql2\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:51 crc kubenswrapper[4948]: I0220 08:20:51.954732 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-lib\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056417 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn4rx\" (UniqueName: \"kubernetes.io/projected/c82772b6-9737-45ee-998a-e897086c03b5-kube-api-access-hn4rx\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056488 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-etc-ovs\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056542 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5ql2\" (UniqueName: \"kubernetes.io/projected/b80cb988-de2a-4e65-9161-a0af0561c754-kube-api-access-h5ql2\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056579 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-lib\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056607 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-log\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056635 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b80cb988-de2a-4e65-9161-a0af0561c754-ovn-controller-tls-certs\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056660 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-run\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056707 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80cb988-de2a-4e65-9161-a0af0561c754-combined-ca-bundle\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056748 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-run\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056778 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-log-ovn\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056799 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-run-ovn\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056824 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b80cb988-de2a-4e65-9161-a0af0561c754-scripts\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.056845 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c82772b6-9737-45ee-998a-e897086c03b5-scripts\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.059251 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c82772b6-9737-45ee-998a-e897086c03b5-scripts\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.059965 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-etc-ovs\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.060345 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-lib\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.060478 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-log\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.061113 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-run\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.061146 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c82772b6-9737-45ee-998a-e897086c03b5-var-run\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.061188 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-run-ovn\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.061228 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b80cb988-de2a-4e65-9161-a0af0561c754-var-log-ovn\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.062929 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b80cb988-de2a-4e65-9161-a0af0561c754-scripts\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.064250 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b80cb988-de2a-4e65-9161-a0af0561c754-ovn-controller-tls-certs\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.066568 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80cb988-de2a-4e65-9161-a0af0561c754-combined-ca-bundle\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.075223 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn4rx\" (UniqueName: \"kubernetes.io/projected/c82772b6-9737-45ee-998a-e897086c03b5-kube-api-access-hn4rx\") pod \"ovn-controller-ovs-q4bv4\" (UID: \"c82772b6-9737-45ee-998a-e897086c03b5\") " pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.079505 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5ql2\" (UniqueName: \"kubernetes.io/projected/b80cb988-de2a-4e65-9161-a0af0561c754-kube-api-access-h5ql2\") pod \"ovn-controller-mkpng\" (UID: \"b80cb988-de2a-4e65-9161-a0af0561c754\") " pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.171128 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mkpng" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.186655 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.731534 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.732855 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.739938 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.740367 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.740666 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-lz8q6" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.740828 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.740910 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.745964 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.869209 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d681442e-6bdb-46e4-9b49-6466e5e036a0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.869271 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.869294 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.869311 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.869336 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.869377 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d681442e-6bdb-46e4-9b49-6466e5e036a0-config\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.869391 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjb6g\" (UniqueName: \"kubernetes.io/projected/d681442e-6bdb-46e4-9b49-6466e5e036a0-kube-api-access-bjb6g\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.869444 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d681442e-6bdb-46e4-9b49-6466e5e036a0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.971283 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d681442e-6bdb-46e4-9b49-6466e5e036a0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.971340 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d681442e-6bdb-46e4-9b49-6466e5e036a0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.971376 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.971391 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.971410 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.971457 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.971495 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d681442e-6bdb-46e4-9b49-6466e5e036a0-config\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.971515 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjb6g\" (UniqueName: \"kubernetes.io/projected/d681442e-6bdb-46e4-9b49-6466e5e036a0-kube-api-access-bjb6g\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.973073 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d681442e-6bdb-46e4-9b49-6466e5e036a0-config\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.973126 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.973525 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d681442e-6bdb-46e4-9b49-6466e5e036a0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.974618 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d681442e-6bdb-46e4-9b49-6466e5e036a0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.979327 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.981747 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:52 crc kubenswrapper[4948]: I0220 08:20:52.982318 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d681442e-6bdb-46e4-9b49-6466e5e036a0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:53 crc kubenswrapper[4948]: I0220 08:20:53.000910 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjb6g\" (UniqueName: \"kubernetes.io/projected/d681442e-6bdb-46e4-9b49-6466e5e036a0-kube-api-access-bjb6g\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:53 crc kubenswrapper[4948]: I0220 08:20:53.015098 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d681442e-6bdb-46e4-9b49-6466e5e036a0\") " pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:53 crc kubenswrapper[4948]: I0220 08:20:53.051331 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.849701 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.851516 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.854528 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.855070 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.855120 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.855159 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-2kzt2" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.861662 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.929682 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf8w4\" (UniqueName: \"kubernetes.io/projected/4b04e19d-cd3d-474d-884d-df12d0e1fffd-kube-api-access-wf8w4\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.929748 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b04e19d-cd3d-474d-884d-df12d0e1fffd-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.929769 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.929791 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.929818 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.930049 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4b04e19d-cd3d-474d-884d-df12d0e1fffd-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.930088 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:55 crc kubenswrapper[4948]: I0220 08:20:55.931077 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b04e19d-cd3d-474d-884d-df12d0e1fffd-config\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.033415 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf8w4\" (UniqueName: \"kubernetes.io/projected/4b04e19d-cd3d-474d-884d-df12d0e1fffd-kube-api-access-wf8w4\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.033657 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b04e19d-cd3d-474d-884d-df12d0e1fffd-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.033688 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.033761 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.033827 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.033872 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4b04e19d-cd3d-474d-884d-df12d0e1fffd-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.033918 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.033951 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b04e19d-cd3d-474d-884d-df12d0e1fffd-config\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.034289 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.034907 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b04e19d-cd3d-474d-884d-df12d0e1fffd-config\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.035050 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4b04e19d-cd3d-474d-884d-df12d0e1fffd-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.035079 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b04e19d-cd3d-474d-884d-df12d0e1fffd-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.041071 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.041247 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.048637 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b04e19d-cd3d-474d-884d-df12d0e1fffd-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.053326 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf8w4\" (UniqueName: \"kubernetes.io/projected/4b04e19d-cd3d-474d-884d-df12d0e1fffd-kube-api-access-wf8w4\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.065128 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4b04e19d-cd3d-474d-884d-df12d0e1fffd\") " pod="openstack/ovsdbserver-sb-0" Feb 20 08:20:56 crc kubenswrapper[4948]: I0220 08:20:56.180926 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 20 08:21:01 crc kubenswrapper[4948]: E0220 08:21:01.610934 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 20 08:21:01 crc kubenswrapper[4948]: E0220 08:21:01.612166 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hdnm9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-mkkzf_openstack(5686ae39-4e44-4303-9b44-ad4b9ceae1ba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:21:01 crc kubenswrapper[4948]: E0220 08:21:01.613284 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.179064 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.796067 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.796880 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pds7s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-zt6gg_openstack(977992e9-30ca-4420-98ae-92f91545cd10): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.798241 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" podUID="977992e9-30ca-4420-98ae-92f91545cd10" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.832203 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.832341 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ffpsp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-cqgz2_openstack(05b1d586-95db-4795-a032-f22ba90e6c16): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.833931 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" podUID="05b1d586-95db-4795-a032-f22ba90e6c16" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.876455 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.876617 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7d754,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-9dfhs_openstack(1224cf30-813c-48f2-92c9-fabe78bbc6bc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:21:02 crc kubenswrapper[4948]: E0220 08:21:02.880070 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" podUID="1224cf30-813c-48f2-92c9-fabe78bbc6bc" Feb 20 08:21:03 crc kubenswrapper[4948]: E0220 08:21:03.186267 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" podUID="977992e9-30ca-4420-98ae-92f91545cd10" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.333268 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 20 08:21:03 crc kubenswrapper[4948]: W0220 08:21:03.348135 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e994489_a2cf_4f6d_a00c_98f627ba0e5f.slice/crio-1e7df2ea4e6039fb737b56f6e8e97afb2ce13f39013495a5535bd3b4cf07ba35 WatchSource:0}: Error finding container 1e7df2ea4e6039fb737b56f6e8e97afb2ce13f39013495a5535bd3b4cf07ba35: Status 404 returned error can't find the container with id 1e7df2ea4e6039fb737b56f6e8e97afb2ce13f39013495a5535bd3b4cf07ba35 Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.359579 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:21:03 crc kubenswrapper[4948]: W0220 08:21:03.371115 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod790894c9_053c_497c_955d_ce7519111dd6.slice/crio-39af94a050f39ff7ebfa915fc855416f40bbca9d8ae91ba23049ec9be5e37b76 WatchSource:0}: Error finding container 39af94a050f39ff7ebfa915fc855416f40bbca9d8ae91ba23049ec9be5e37b76: Status 404 returned error can't find the container with id 39af94a050f39ff7ebfa915fc855416f40bbca9d8ae91ba23049ec9be5e37b76 Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.513939 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.529786 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.584619 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-dns-svc\") pod \"05b1d586-95db-4795-a032-f22ba90e6c16\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.584710 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-config\") pod \"05b1d586-95db-4795-a032-f22ba90e6c16\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.584791 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffpsp\" (UniqueName: \"kubernetes.io/projected/05b1d586-95db-4795-a032-f22ba90e6c16-kube-api-access-ffpsp\") pod \"05b1d586-95db-4795-a032-f22ba90e6c16\" (UID: \"05b1d586-95db-4795-a032-f22ba90e6c16\") " Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.585322 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "05b1d586-95db-4795-a032-f22ba90e6c16" (UID: "05b1d586-95db-4795-a032-f22ba90e6c16"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.585512 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-config" (OuterVolumeSpecName: "config") pod "05b1d586-95db-4795-a032-f22ba90e6c16" (UID: "05b1d586-95db-4795-a032-f22ba90e6c16"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.588863 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05b1d586-95db-4795-a032-f22ba90e6c16-kube-api-access-ffpsp" (OuterVolumeSpecName: "kube-api-access-ffpsp") pod "05b1d586-95db-4795-a032-f22ba90e6c16" (UID: "05b1d586-95db-4795-a032-f22ba90e6c16"). InnerVolumeSpecName "kube-api-access-ffpsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.634104 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rt8tp"] Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.649591 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.658492 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.664614 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mkpng"] Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.670588 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.685865 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1224cf30-813c-48f2-92c9-fabe78bbc6bc-config\") pod \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\" (UID: \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\") " Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.685901 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7d754\" (UniqueName: \"kubernetes.io/projected/1224cf30-813c-48f2-92c9-fabe78bbc6bc-kube-api-access-7d754\") pod \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\" (UID: \"1224cf30-813c-48f2-92c9-fabe78bbc6bc\") " Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.686292 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffpsp\" (UniqueName: \"kubernetes.io/projected/05b1d586-95db-4795-a032-f22ba90e6c16-kube-api-access-ffpsp\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.686311 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.686322 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05b1d586-95db-4795-a032-f22ba90e6c16-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.687184 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1224cf30-813c-48f2-92c9-fabe78bbc6bc-config" (OuterVolumeSpecName: "config") pod "1224cf30-813c-48f2-92c9-fabe78bbc6bc" (UID: "1224cf30-813c-48f2-92c9-fabe78bbc6bc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.734937 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1224cf30-813c-48f2-92c9-fabe78bbc6bc-kube-api-access-7d754" (OuterVolumeSpecName: "kube-api-access-7d754") pod "1224cf30-813c-48f2-92c9-fabe78bbc6bc" (UID: "1224cf30-813c-48f2-92c9-fabe78bbc6bc"). InnerVolumeSpecName "kube-api-access-7d754". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.786620 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-q4bv4"] Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.787833 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1224cf30-813c-48f2-92c9-fabe78bbc6bc-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:03 crc kubenswrapper[4948]: I0220 08:21:03.787870 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7d754\" (UniqueName: \"kubernetes.io/projected/1224cf30-813c-48f2-92c9-fabe78bbc6bc-kube-api-access-7d754\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:03 crc kubenswrapper[4948]: W0220 08:21:03.790500 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0dfbb200_e444_4895_a410_e51cbb1fc112.slice/crio-8d55ac91670695a37a460e364e6527e261fc751cd8cb9dd56d60bf1aa04e7514 WatchSource:0}: Error finding container 8d55ac91670695a37a460e364e6527e261fc751cd8cb9dd56d60bf1aa04e7514: Status 404 returned error can't find the container with id 8d55ac91670695a37a460e364e6527e261fc751cd8cb9dd56d60bf1aa04e7514 Feb 20 08:21:03 crc kubenswrapper[4948]: W0220 08:21:03.801444 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc82772b6_9737_45ee_998a_e897086c03b5.slice/crio-ec18154bf98d02691b4396af19785093698c4033031de11ba963723e7ece9f00 WatchSource:0}: Error finding container ec18154bf98d02691b4396af19785093698c4033031de11ba963723e7ece9f00: Status 404 returned error can't find the container with id ec18154bf98d02691b4396af19785093698c4033031de11ba963723e7ece9f00 Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.190623 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" event={"ID":"1224cf30-813c-48f2-92c9-fabe78bbc6bc","Type":"ContainerDied","Data":"459736be569ebf9acfc7e517d56782883ab07764b913b8c4d305735259e95b55"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.190723 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9dfhs" Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.192885 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0dfbb200-e444-4895-a410-e51cbb1fc112","Type":"ContainerStarted","Data":"8d55ac91670695a37a460e364e6527e261fc751cd8cb9dd56d60bf1aa04e7514"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.197286 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"73b84bb7-f594-4823-ac03-40fdac6ee177","Type":"ContainerStarted","Data":"d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.207417 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"790894c9-053c-497c-955d-ce7519111dd6","Type":"ContainerStarted","Data":"39af94a050f39ff7ebfa915fc855416f40bbca9d8ae91ba23049ec9be5e37b76"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.208741 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d02114ae-21f2-4cc5-9bca-9505c25eaf5f","Type":"ContainerStarted","Data":"88f7820770eaae07d0058f62cddfa17073943d7aca489b897da76d305b4a45a0"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.210430 4948 generic.go:334] "Generic (PLEG): container finished" podID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerID="b689fa43e3b5c02fa357d00443dc58a7d6fa6d71f6692b435028801366e8a5e6" exitCode=0 Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.210489 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt8tp" event={"ID":"93f24735-d27d-4867-b9f7-cce47e7f376f","Type":"ContainerDied","Data":"b689fa43e3b5c02fa357d00443dc58a7d6fa6d71f6692b435028801366e8a5e6"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.210511 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt8tp" event={"ID":"93f24735-d27d-4867-b9f7-cce47e7f376f","Type":"ContainerStarted","Data":"104e51617975f8f0d6bf864e106e2486e0a18298cc9a24f078002083a6d423b2"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.212267 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mkpng" event={"ID":"b80cb988-de2a-4e65-9161-a0af0561c754","Type":"ContainerStarted","Data":"128bc4be043ddeeae350cc9c95a3938b26679c4935fdd074d497449c5afcb7a9"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.213323 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"8e994489-a2cf-4f6d-a00c-98f627ba0e5f","Type":"ContainerStarted","Data":"1e7df2ea4e6039fb737b56f6e8e97afb2ce13f39013495a5535bd3b4cf07ba35"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.217199 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" event={"ID":"05b1d586-95db-4795-a032-f22ba90e6c16","Type":"ContainerDied","Data":"4e5e167063afcb1f5336b93cee715fd831f959a4a5fb12e9e562274ec18ad672"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.217322 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-cqgz2" Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.247376 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6e59b63d-0c56-488c-87cb-348af87058c6","Type":"ContainerStarted","Data":"0391c43404c3571e63693b407b5d3441d92363fdd709b3395357f5dcce9c9f30"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.249568 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q4bv4" event={"ID":"c82772b6-9737-45ee-998a-e897086c03b5","Type":"ContainerStarted","Data":"ec18154bf98d02691b4396af19785093698c4033031de11ba963723e7ece9f00"} Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.262656 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9dfhs"] Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.265640 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9dfhs"] Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.292373 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cqgz2"] Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.297273 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cqgz2"] Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.350716 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 20 08:21:04 crc kubenswrapper[4948]: W0220 08:21:04.357245 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd681442e_6bdb_46e4_9b49_6466e5e036a0.slice/crio-d8ed751d1b95575a7a1d2d4ba48699ab1ced21c222a4164eb4cdb6f411745130 WatchSource:0}: Error finding container d8ed751d1b95575a7a1d2d4ba48699ab1ced21c222a4164eb4cdb6f411745130: Status 404 returned error can't find the container with id d8ed751d1b95575a7a1d2d4ba48699ab1ced21c222a4164eb4cdb6f411745130 Feb 20 08:21:04 crc kubenswrapper[4948]: I0220 08:21:04.585935 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 20 08:21:04 crc kubenswrapper[4948]: W0220 08:21:04.923943 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b04e19d_cd3d_474d_884d_df12d0e1fffd.slice/crio-660ea7b70688ba8b37a938c6b7d318fb1a1c77090b2c1e58092a4070f0016d7c WatchSource:0}: Error finding container 660ea7b70688ba8b37a938c6b7d318fb1a1c77090b2c1e58092a4070f0016d7c: Status 404 returned error can't find the container with id 660ea7b70688ba8b37a938c6b7d318fb1a1c77090b2c1e58092a4070f0016d7c Feb 20 08:21:05 crc kubenswrapper[4948]: I0220 08:21:05.257388 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"790894c9-053c-497c-955d-ce7519111dd6","Type":"ContainerStarted","Data":"77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242"} Feb 20 08:21:05 crc kubenswrapper[4948]: I0220 08:21:05.260137 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d681442e-6bdb-46e4-9b49-6466e5e036a0","Type":"ContainerStarted","Data":"d8ed751d1b95575a7a1d2d4ba48699ab1ced21c222a4164eb4cdb6f411745130"} Feb 20 08:21:05 crc kubenswrapper[4948]: I0220 08:21:05.262086 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4b04e19d-cd3d-474d-884d-df12d0e1fffd","Type":"ContainerStarted","Data":"660ea7b70688ba8b37a938c6b7d318fb1a1c77090b2c1e58092a4070f0016d7c"} Feb 20 08:21:06 crc kubenswrapper[4948]: I0220 08:21:05.737536 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05b1d586-95db-4795-a032-f22ba90e6c16" path="/var/lib/kubelet/pods/05b1d586-95db-4795-a032-f22ba90e6c16/volumes" Feb 20 08:21:06 crc kubenswrapper[4948]: I0220 08:21:05.738726 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1224cf30-813c-48f2-92c9-fabe78bbc6bc" path="/var/lib/kubelet/pods/1224cf30-813c-48f2-92c9-fabe78bbc6bc/volumes" Feb 20 08:21:08 crc kubenswrapper[4948]: I0220 08:21:08.025005 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:21:08 crc kubenswrapper[4948]: I0220 08:21:08.025498 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.365859 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6e59b63d-0c56-488c-87cb-348af87058c6","Type":"ContainerStarted","Data":"52aab5e3e31e0e0c819bc56b6e3130d428a9243c172c9a352f49b7f3ba0db78d"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.368167 4948 generic.go:334] "Generic (PLEG): container finished" podID="c82772b6-9737-45ee-998a-e897086c03b5" containerID="ff1b1e67637ea3148c6b86c285290aa88c29f45d19a025c92565c44923597b25" exitCode=0 Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.368222 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q4bv4" event={"ID":"c82772b6-9737-45ee-998a-e897086c03b5","Type":"ContainerDied","Data":"ff1b1e67637ea3148c6b86c285290aa88c29f45d19a025c92565c44923597b25"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.372242 4948 generic.go:334] "Generic (PLEG): container finished" podID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerID="7f925a093fb8a5b68097ec9940401bb5483d01b61401a6a70fda1f29efe6615e" exitCode=0 Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.372280 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt8tp" event={"ID":"93f24735-d27d-4867-b9f7-cce47e7f376f","Type":"ContainerDied","Data":"7f925a093fb8a5b68097ec9940401bb5483d01b61401a6a70fda1f29efe6615e"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.375792 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d02114ae-21f2-4cc5-9bca-9505c25eaf5f","Type":"ContainerStarted","Data":"3ef0ae79740a0d142692c9fc559c45d8a8e3955ca475b2d1eb3cc2ae8aea4eac"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.377904 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"8e994489-a2cf-4f6d-a00c-98f627ba0e5f","Type":"ContainerStarted","Data":"6209e5b7193825550a19ad6f82db51db28926334103fbbb1ba1accdf4480217a"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.378350 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.385275 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4b04e19d-cd3d-474d-884d-df12d0e1fffd","Type":"ContainerStarted","Data":"2982b3c30a182ecd9b19a054cd88fa344147eba3d2ef83da730f1b6236bef02e"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.387960 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0dfbb200-e444-4895-a410-e51cbb1fc112","Type":"ContainerStarted","Data":"20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.388842 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.391303 4948 generic.go:334] "Generic (PLEG): container finished" podID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" containerID="5a5bfff9ba7b803c56d99738536efa4fb27865dfba1fd58a67dc23b1d4bef9ed" exitCode=0 Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.391380 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" event={"ID":"5686ae39-4e44-4303-9b44-ad4b9ceae1ba","Type":"ContainerDied","Data":"5a5bfff9ba7b803c56d99738536efa4fb27865dfba1fd58a67dc23b1d4bef9ed"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.404309 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mkpng" event={"ID":"b80cb988-de2a-4e65-9161-a0af0561c754","Type":"ContainerStarted","Data":"99d11716eb57476f8ffaf8e1e981cf08ec5515e1fcf0b7d0d1ca6915191b7c59"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.404472 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-mkpng" Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.409795 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d681442e-6bdb-46e4-9b49-6466e5e036a0","Type":"ContainerStarted","Data":"154b3b91cde6adf519be9c377c4fa145eb9c9b5eb16cfea11a3489cc91629ba5"} Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.429330 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=19.895896955 podStartE2EDuration="30.429316495s" podCreationTimestamp="2026-02-20 08:20:46 +0000 UTC" firstStartedPulling="2026-02-20 08:21:03.357149088 +0000 UTC m=+912.331643928" lastFinishedPulling="2026-02-20 08:21:13.890568638 +0000 UTC m=+922.865063468" observedRunningTime="2026-02-20 08:21:16.428549546 +0000 UTC m=+925.403044366" watchObservedRunningTime="2026-02-20 08:21:16.429316495 +0000 UTC m=+925.403811305" Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.495931 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=15.880916618 podStartE2EDuration="27.495908993s" podCreationTimestamp="2026-02-20 08:20:49 +0000 UTC" firstStartedPulling="2026-02-20 08:21:03.795894026 +0000 UTC m=+912.770388846" lastFinishedPulling="2026-02-20 08:21:15.410886401 +0000 UTC m=+924.385381221" observedRunningTime="2026-02-20 08:21:16.481500036 +0000 UTC m=+925.455994856" watchObservedRunningTime="2026-02-20 08:21:16.495908993 +0000 UTC m=+925.470403813" Feb 20 08:21:16 crc kubenswrapper[4948]: I0220 08:21:16.502306 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-mkpng" podStartSLOduration=14.949192415 podStartE2EDuration="25.502290311s" podCreationTimestamp="2026-02-20 08:20:51 +0000 UTC" firstStartedPulling="2026-02-20 08:21:03.76614449 +0000 UTC m=+912.740639310" lastFinishedPulling="2026-02-20 08:21:14.319242386 +0000 UTC m=+923.293737206" observedRunningTime="2026-02-20 08:21:16.497633985 +0000 UTC m=+925.472128815" watchObservedRunningTime="2026-02-20 08:21:16.502290311 +0000 UTC m=+925.476785131" Feb 20 08:21:17 crc kubenswrapper[4948]: I0220 08:21:17.420666 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" event={"ID":"5686ae39-4e44-4303-9b44-ad4b9ceae1ba","Type":"ContainerStarted","Data":"f8a24951a6fecaa760fab34934fbaaebc5ce586261d093c015219f82cccd0422"} Feb 20 08:21:17 crc kubenswrapper[4948]: I0220 08:21:17.421379 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:21:17 crc kubenswrapper[4948]: I0220 08:21:17.425398 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q4bv4" event={"ID":"c82772b6-9737-45ee-998a-e897086c03b5","Type":"ContainerStarted","Data":"52a4b050691a7d6574abd457527106d71c476df878106b33767014ec3321039e"} Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.435292 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4b04e19d-cd3d-474d-884d-df12d0e1fffd","Type":"ContainerStarted","Data":"2e87e5a50286fc9148333fb04efebbd5fb4d85acaf989d4e9a4d48c3df249122"} Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.437569 4948 generic.go:334] "Generic (PLEG): container finished" podID="977992e9-30ca-4420-98ae-92f91545cd10" containerID="05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099" exitCode=0 Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.437647 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" event={"ID":"977992e9-30ca-4420-98ae-92f91545cd10","Type":"ContainerDied","Data":"05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099"} Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.440734 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q4bv4" event={"ID":"c82772b6-9737-45ee-998a-e897086c03b5","Type":"ContainerStarted","Data":"45d366f5451bce0eea4471ff0fa712033b40558f9b4867b32e46492ba2b78eaa"} Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.440910 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.444561 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt8tp" event={"ID":"93f24735-d27d-4867-b9f7-cce47e7f376f","Type":"ContainerStarted","Data":"fbf3b4c07c31284ef058431c15f6fb31624a27a446740ecaf65fcf08b863a6e3"} Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.447952 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d681442e-6bdb-46e4-9b49-6466e5e036a0","Type":"ContainerStarted","Data":"c93fa988182cb818cfb4eb867eabf542e0d664bbe1a2420a5257aa6e4180960f"} Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.483150 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=11.916423383 podStartE2EDuration="24.48312883s" podCreationTimestamp="2026-02-20 08:20:54 +0000 UTC" firstStartedPulling="2026-02-20 08:21:04.935089118 +0000 UTC m=+913.909583978" lastFinishedPulling="2026-02-20 08:21:17.501794605 +0000 UTC m=+926.476289425" observedRunningTime="2026-02-20 08:21:18.481406507 +0000 UTC m=+927.455901327" watchObservedRunningTime="2026-02-20 08:21:18.48312883 +0000 UTC m=+927.457623660" Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.484382 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" podStartSLOduration=5.125547855 podStartE2EDuration="36.484372111s" podCreationTimestamp="2026-02-20 08:20:42 +0000 UTC" firstStartedPulling="2026-02-20 08:20:43.884879158 +0000 UTC m=+892.859373978" lastFinishedPulling="2026-02-20 08:21:15.243703394 +0000 UTC m=+924.218198234" observedRunningTime="2026-02-20 08:21:17.441761979 +0000 UTC m=+926.416256809" watchObservedRunningTime="2026-02-20 08:21:18.484372111 +0000 UTC m=+927.458866941" Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.539493 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-q4bv4" podStartSLOduration=17.454977274 podStartE2EDuration="27.539476124s" podCreationTimestamp="2026-02-20 08:20:51 +0000 UTC" firstStartedPulling="2026-02-20 08:21:03.804730965 +0000 UTC m=+912.779225785" lastFinishedPulling="2026-02-20 08:21:13.889229805 +0000 UTC m=+922.863724635" observedRunningTime="2026-02-20 08:21:18.535343732 +0000 UTC m=+927.509838552" watchObservedRunningTime="2026-02-20 08:21:18.539476124 +0000 UTC m=+927.513970944" Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.601658 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=14.463602167 podStartE2EDuration="27.601640393s" podCreationTimestamp="2026-02-20 08:20:51 +0000 UTC" firstStartedPulling="2026-02-20 08:21:04.36380509 +0000 UTC m=+913.338299920" lastFinishedPulling="2026-02-20 08:21:17.501843316 +0000 UTC m=+926.476338146" observedRunningTime="2026-02-20 08:21:18.573856265 +0000 UTC m=+927.548351075" watchObservedRunningTime="2026-02-20 08:21:18.601640393 +0000 UTC m=+927.576135213" Feb 20 08:21:18 crc kubenswrapper[4948]: I0220 08:21:18.703157 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rt8tp" podStartSLOduration=20.417332932 podStartE2EDuration="33.703139445s" podCreationTimestamp="2026-02-20 08:20:45 +0000 UTC" firstStartedPulling="2026-02-20 08:21:04.213406028 +0000 UTC m=+913.187900858" lastFinishedPulling="2026-02-20 08:21:17.499212551 +0000 UTC m=+926.473707371" observedRunningTime="2026-02-20 08:21:18.692630885 +0000 UTC m=+927.667125695" watchObservedRunningTime="2026-02-20 08:21:18.703139445 +0000 UTC m=+927.677634265" Feb 20 08:21:19 crc kubenswrapper[4948]: I0220 08:21:19.459460 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" event={"ID":"977992e9-30ca-4420-98ae-92f91545cd10","Type":"ContainerStarted","Data":"2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a"} Feb 20 08:21:19 crc kubenswrapper[4948]: I0220 08:21:19.460885 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:21:19 crc kubenswrapper[4948]: I0220 08:21:19.461482 4948 generic.go:334] "Generic (PLEG): container finished" podID="6e59b63d-0c56-488c-87cb-348af87058c6" containerID="52aab5e3e31e0e0c819bc56b6e3130d428a9243c172c9a352f49b7f3ba0db78d" exitCode=0 Feb 20 08:21:19 crc kubenswrapper[4948]: I0220 08:21:19.461575 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6e59b63d-0c56-488c-87cb-348af87058c6","Type":"ContainerDied","Data":"52aab5e3e31e0e0c819bc56b6e3130d428a9243c172c9a352f49b7f3ba0db78d"} Feb 20 08:21:19 crc kubenswrapper[4948]: I0220 08:21:19.462928 4948 generic.go:334] "Generic (PLEG): container finished" podID="d02114ae-21f2-4cc5-9bca-9505c25eaf5f" containerID="3ef0ae79740a0d142692c9fc559c45d8a8e3955ca475b2d1eb3cc2ae8aea4eac" exitCode=0 Feb 20 08:21:19 crc kubenswrapper[4948]: I0220 08:21:19.463141 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d02114ae-21f2-4cc5-9bca-9505c25eaf5f","Type":"ContainerDied","Data":"3ef0ae79740a0d142692c9fc559c45d8a8e3955ca475b2d1eb3cc2ae8aea4eac"} Feb 20 08:21:19 crc kubenswrapper[4948]: I0220 08:21:19.464834 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:21:19 crc kubenswrapper[4948]: I0220 08:21:19.494654 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" podStartSLOduration=-9223371999.360136 podStartE2EDuration="37.494639802s" podCreationTimestamp="2026-02-20 08:20:42 +0000 UTC" firstStartedPulling="2026-02-20 08:20:43.472030076 +0000 UTC m=+892.446524896" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:19.493017722 +0000 UTC m=+928.467512582" watchObservedRunningTime="2026-02-20 08:21:19.494639802 +0000 UTC m=+928.469134622" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.052900 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.104227 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.182126 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.225918 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.477744 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d02114ae-21f2-4cc5-9bca-9505c25eaf5f","Type":"ContainerStarted","Data":"71988e03aff6de789cff2b8a78a3ac5d432604d0dc5f50a7f67a3f2ee9dd935d"} Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.482415 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6e59b63d-0c56-488c-87cb-348af87058c6","Type":"ContainerStarted","Data":"c46f534c65704b9778549e3e9e455d8942dd90bd2303681319fcb6071803ed28"} Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.483573 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.483609 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.511433 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=25.376815015 podStartE2EDuration="35.511417725s" podCreationTimestamp="2026-02-20 08:20:45 +0000 UTC" firstStartedPulling="2026-02-20 08:21:03.754600284 +0000 UTC m=+912.729095104" lastFinishedPulling="2026-02-20 08:21:13.889202994 +0000 UTC m=+922.863697814" observedRunningTime="2026-02-20 08:21:20.509831225 +0000 UTC m=+929.484326045" watchObservedRunningTime="2026-02-20 08:21:20.511417725 +0000 UTC m=+929.485912545" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.544375 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=25.895116594 podStartE2EDuration="36.54435095s" podCreationTimestamp="2026-02-20 08:20:44 +0000 UTC" firstStartedPulling="2026-02-20 08:21:03.754561803 +0000 UTC m=+912.729056623" lastFinishedPulling="2026-02-20 08:21:14.403796149 +0000 UTC m=+923.378290979" observedRunningTime="2026-02-20 08:21:20.537819448 +0000 UTC m=+929.512314278" watchObservedRunningTime="2026-02-20 08:21:20.54435095 +0000 UTC m=+929.518845780" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.552127 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.555522 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.721913 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mkkzf"] Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.723084 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" containerName="dnsmasq-dns" containerID="cri-o://f8a24951a6fecaa760fab34934fbaaebc5ce586261d093c015219f82cccd0422" gracePeriod=10 Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.774601 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-mvb6f"] Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.776106 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.778073 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.781224 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-gcfn8"] Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.784095 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.785421 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.849286 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-gcfn8"] Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858502 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1510e8b2-af6f-4641-a2d2-361830382250-config\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858583 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cpkv\" (UniqueName: \"kubernetes.io/projected/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-kube-api-access-9cpkv\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858602 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858626 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1510e8b2-af6f-4641-a2d2-361830382250-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858653 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw495\" (UniqueName: \"kubernetes.io/projected/1510e8b2-af6f-4641-a2d2-361830382250-kube-api-access-kw495\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858673 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1510e8b2-af6f-4641-a2d2-361830382250-ovs-rundir\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858725 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1510e8b2-af6f-4641-a2d2-361830382250-combined-ca-bundle\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858739 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1510e8b2-af6f-4641-a2d2-361830382250-ovn-rundir\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858758 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.858784 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.878432 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-mvb6f"] Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.914324 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zt6gg"] Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.949310 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hmhzl"] Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.950550 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.960449 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961490 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961543 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1510e8b2-af6f-4641-a2d2-361830382250-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961575 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw495\" (UniqueName: \"kubernetes.io/projected/1510e8b2-af6f-4641-a2d2-361830382250-kube-api-access-kw495\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961603 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1510e8b2-af6f-4641-a2d2-361830382250-ovs-rundir\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961649 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1510e8b2-af6f-4641-a2d2-361830382250-combined-ca-bundle\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961664 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1510e8b2-af6f-4641-a2d2-361830382250-ovn-rundir\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961687 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961717 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961768 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1510e8b2-af6f-4641-a2d2-361830382250-config\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.961820 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cpkv\" (UniqueName: \"kubernetes.io/projected/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-kube-api-access-9cpkv\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.963220 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1510e8b2-af6f-4641-a2d2-361830382250-ovn-rundir\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.964755 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1510e8b2-af6f-4641-a2d2-361830382250-ovs-rundir\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.964984 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.965399 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.965546 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.965668 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1510e8b2-af6f-4641-a2d2-361830382250-config\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.972059 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1510e8b2-af6f-4641-a2d2-361830382250-combined-ca-bundle\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.975275 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1510e8b2-af6f-4641-a2d2-361830382250-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.976776 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hmhzl"] Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.987115 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cpkv\" (UniqueName: \"kubernetes.io/projected/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-kube-api-access-9cpkv\") pod \"dnsmasq-dns-7fd796d7df-mvb6f\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:20 crc kubenswrapper[4948]: I0220 08:21:20.989050 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw495\" (UniqueName: \"kubernetes.io/projected/1510e8b2-af6f-4641-a2d2-361830382250-kube-api-access-kw495\") pod \"ovn-controller-metrics-gcfn8\" (UID: \"1510e8b2-af6f-4641-a2d2-361830382250\") " pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.004414 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.006361 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.010914 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.019195 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-bpq2x" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.019348 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.019469 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.019651 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.063841 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/eeb3782c-0a88-4d17-b481-e2aef7ea386a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.063955 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-config\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064035 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064086 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064192 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064227 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7h4r\" (UniqueName: \"kubernetes.io/projected/eeb3782c-0a88-4d17-b481-e2aef7ea386a-kube-api-access-c7h4r\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064279 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eeb3782c-0a88-4d17-b481-e2aef7ea386a-scripts\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064328 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064468 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeb3782c-0a88-4d17-b481-e2aef7ea386a-config\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064490 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064506 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2twp\" (UniqueName: \"kubernetes.io/projected/33a06f41-ea8b-4138-8048-5425cd04f015-kube-api-access-c2twp\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.064663 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166452 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166497 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166544 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166565 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7h4r\" (UniqueName: \"kubernetes.io/projected/eeb3782c-0a88-4d17-b481-e2aef7ea386a-kube-api-access-c7h4r\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166599 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eeb3782c-0a88-4d17-b481-e2aef7ea386a-scripts\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166628 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166659 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166679 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2twp\" (UniqueName: \"kubernetes.io/projected/33a06f41-ea8b-4138-8048-5425cd04f015-kube-api-access-c2twp\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166700 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeb3782c-0a88-4d17-b481-e2aef7ea386a-config\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166733 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166763 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/eeb3782c-0a88-4d17-b481-e2aef7ea386a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.166823 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-config\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.168083 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-config\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.169060 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/eeb3782c-0a88-4d17-b481-e2aef7ea386a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.169319 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.171877 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.172580 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.173020 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eeb3782c-0a88-4d17-b481-e2aef7ea386a-scripts\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.173193 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.173378 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.173650 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb3782c-0a88-4d17-b481-e2aef7ea386a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.174080 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeb3782c-0a88-4d17-b481-e2aef7ea386a-config\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.185736 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2twp\" (UniqueName: \"kubernetes.io/projected/33a06f41-ea8b-4138-8048-5425cd04f015-kube-api-access-c2twp\") pod \"dnsmasq-dns-86db49b7ff-hmhzl\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.188430 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7h4r\" (UniqueName: \"kubernetes.io/projected/eeb3782c-0a88-4d17-b481-e2aef7ea386a-kube-api-access-c7h4r\") pod \"ovn-northd-0\" (UID: \"eeb3782c-0a88-4d17-b481-e2aef7ea386a\") " pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.191788 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.199018 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-gcfn8" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.374860 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.388830 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.494404 4948 generic.go:334] "Generic (PLEG): container finished" podID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" containerID="f8a24951a6fecaa760fab34934fbaaebc5ce586261d093c015219f82cccd0422" exitCode=0 Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.496097 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" event={"ID":"5686ae39-4e44-4303-9b44-ad4b9ceae1ba","Type":"ContainerDied","Data":"f8a24951a6fecaa760fab34934fbaaebc5ce586261d093c015219f82cccd0422"} Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.496311 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" podUID="977992e9-30ca-4420-98ae-92f91545cd10" containerName="dnsmasq-dns" containerID="cri-o://2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a" gracePeriod=10 Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.793470 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-mvb6f"] Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.819101 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-gcfn8"] Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.845495 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hmhzl"] Feb 20 08:21:21 crc kubenswrapper[4948]: I0220 08:21:21.854647 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.143491 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.148113 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.195463 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdnm9\" (UniqueName: \"kubernetes.io/projected/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-kube-api-access-hdnm9\") pod \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.195548 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-dns-svc\") pod \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.195615 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-config\") pod \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\" (UID: \"5686ae39-4e44-4303-9b44-ad4b9ceae1ba\") " Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.202124 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-kube-api-access-hdnm9" (OuterVolumeSpecName: "kube-api-access-hdnm9") pod "5686ae39-4e44-4303-9b44-ad4b9ceae1ba" (UID: "5686ae39-4e44-4303-9b44-ad4b9ceae1ba"). InnerVolumeSpecName "kube-api-access-hdnm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.298314 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdnm9\" (UniqueName: \"kubernetes.io/projected/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-kube-api-access-hdnm9\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.363133 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5686ae39-4e44-4303-9b44-ad4b9ceae1ba" (UID: "5686ae39-4e44-4303-9b44-ad4b9ceae1ba"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.399796 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.420489 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-config" (OuterVolumeSpecName: "config") pod "5686ae39-4e44-4303-9b44-ad4b9ceae1ba" (UID: "5686ae39-4e44-4303-9b44-ad4b9ceae1ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.437466 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.506952 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5686ae39-4e44-4303-9b44-ad4b9ceae1ba-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.519936 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-gcfn8" event={"ID":"1510e8b2-af6f-4641-a2d2-361830382250","Type":"ContainerStarted","Data":"199d2767067990d2fddd48e26caa156fee51020bbc1379e0aa8846fdfc7146d7"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.520000 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-gcfn8" event={"ID":"1510e8b2-af6f-4641-a2d2-361830382250","Type":"ContainerStarted","Data":"1b538d3cc5f8e21b1fd8955b6a0f0b465e33747d8eced9c6d602b588e10366c9"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.522045 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" event={"ID":"5686ae39-4e44-4303-9b44-ad4b9ceae1ba","Type":"ContainerDied","Data":"b4980c083e552022d3673e8c44cd0e8286ee33627cbef5030e75bd7be40133d3"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.522114 4948 scope.go:117] "RemoveContainer" containerID="f8a24951a6fecaa760fab34934fbaaebc5ce586261d093c015219f82cccd0422" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.522258 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mkkzf" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.526191 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"eeb3782c-0a88-4d17-b481-e2aef7ea386a","Type":"ContainerStarted","Data":"3df4a2193c9907b10329446f679b3d045c8e0838be144262dd416a7125f8a215"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.528286 4948 generic.go:334] "Generic (PLEG): container finished" podID="977992e9-30ca-4420-98ae-92f91545cd10" containerID="2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a" exitCode=0 Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.528345 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" event={"ID":"977992e9-30ca-4420-98ae-92f91545cd10","Type":"ContainerDied","Data":"2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.528365 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" event={"ID":"977992e9-30ca-4420-98ae-92f91545cd10","Type":"ContainerDied","Data":"a0508d6a1f4b7f39846aac7f90bdfe83ad8793d4225f8bb5f2eaca0710969807"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.528436 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zt6gg" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.532014 4948 generic.go:334] "Generic (PLEG): container finished" podID="33a06f41-ea8b-4138-8048-5425cd04f015" containerID="664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a" exitCode=0 Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.532073 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" event={"ID":"33a06f41-ea8b-4138-8048-5425cd04f015","Type":"ContainerDied","Data":"664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.532099 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" event={"ID":"33a06f41-ea8b-4138-8048-5425cd04f015","Type":"ContainerStarted","Data":"1f7e5587a47afc9ffa01b7b40dc0afdae69e9922140a4a9e4162b6a2a7417196"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.533699 4948 generic.go:334] "Generic (PLEG): container finished" podID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" containerID="c624ced0b3cd9432f64567b57ff22f472c7a407536a9a515e94a3ab95cf57662" exitCode=0 Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.534770 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" event={"ID":"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be","Type":"ContainerDied","Data":"c624ced0b3cd9432f64567b57ff22f472c7a407536a9a515e94a3ab95cf57662"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.534791 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" event={"ID":"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be","Type":"ContainerStarted","Data":"a0b4a492fe4751057387e9dea455a523858f53f3d884620fffd805ea3be59755"} Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.553252 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-gcfn8" podStartSLOduration=2.553236133 podStartE2EDuration="2.553236133s" podCreationTimestamp="2026-02-20 08:21:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:22.548639479 +0000 UTC m=+931.523134299" watchObservedRunningTime="2026-02-20 08:21:22.553236133 +0000 UTC m=+931.527730953" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.608455 4948 scope.go:117] "RemoveContainer" containerID="5a5bfff9ba7b803c56d99738536efa4fb27865dfba1fd58a67dc23b1d4bef9ed" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.615914 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-config\") pod \"977992e9-30ca-4420-98ae-92f91545cd10\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.616799 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pds7s\" (UniqueName: \"kubernetes.io/projected/977992e9-30ca-4420-98ae-92f91545cd10-kube-api-access-pds7s\") pod \"977992e9-30ca-4420-98ae-92f91545cd10\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.616839 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-dns-svc\") pod \"977992e9-30ca-4420-98ae-92f91545cd10\" (UID: \"977992e9-30ca-4420-98ae-92f91545cd10\") " Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.623757 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/977992e9-30ca-4420-98ae-92f91545cd10-kube-api-access-pds7s" (OuterVolumeSpecName: "kube-api-access-pds7s") pod "977992e9-30ca-4420-98ae-92f91545cd10" (UID: "977992e9-30ca-4420-98ae-92f91545cd10"). InnerVolumeSpecName "kube-api-access-pds7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.672508 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mkkzf"] Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.675393 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "977992e9-30ca-4420-98ae-92f91545cd10" (UID: "977992e9-30ca-4420-98ae-92f91545cd10"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.684064 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-config" (OuterVolumeSpecName: "config") pod "977992e9-30ca-4420-98ae-92f91545cd10" (UID: "977992e9-30ca-4420-98ae-92f91545cd10"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.692460 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mkkzf"] Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.693883 4948 scope.go:117] "RemoveContainer" containerID="2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.719952 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.720004 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977992e9-30ca-4420-98ae-92f91545cd10-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.720014 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pds7s\" (UniqueName: \"kubernetes.io/projected/977992e9-30ca-4420-98ae-92f91545cd10-kube-api-access-pds7s\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.724694 4948 scope.go:117] "RemoveContainer" containerID="05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.755945 4948 scope.go:117] "RemoveContainer" containerID="2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a" Feb 20 08:21:22 crc kubenswrapper[4948]: E0220 08:21:22.756659 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a\": container with ID starting with 2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a not found: ID does not exist" containerID="2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.756685 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a"} err="failed to get container status \"2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a\": rpc error: code = NotFound desc = could not find container \"2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a\": container with ID starting with 2fa62b0646099a9e4d67d43a31c1228c43fee77b21220b1456b925205d581d9a not found: ID does not exist" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.756706 4948 scope.go:117] "RemoveContainer" containerID="05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099" Feb 20 08:21:22 crc kubenswrapper[4948]: E0220 08:21:22.757008 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099\": container with ID starting with 05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099 not found: ID does not exist" containerID="05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.757026 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099"} err="failed to get container status \"05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099\": rpc error: code = NotFound desc = could not find container \"05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099\": container with ID starting with 05ccd3525150ab098e3561398cfda7a7a04a42e4d12ac43eacc78bfa91ca9099 not found: ID does not exist" Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.921241 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zt6gg"] Feb 20 08:21:22 crc kubenswrapper[4948]: I0220 08:21:22.927468 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zt6gg"] Feb 20 08:21:23 crc kubenswrapper[4948]: I0220 08:21:23.546058 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" event={"ID":"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be","Type":"ContainerStarted","Data":"ae779bfaf190f36bd8dce0329a6a4c0ac105d70cb79287b3be92631d581020d5"} Feb 20 08:21:23 crc kubenswrapper[4948]: I0220 08:21:23.546396 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:23 crc kubenswrapper[4948]: I0220 08:21:23.550231 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" event={"ID":"33a06f41-ea8b-4138-8048-5425cd04f015","Type":"ContainerStarted","Data":"489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f"} Feb 20 08:21:23 crc kubenswrapper[4948]: I0220 08:21:23.569174 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" podStartSLOduration=3.569156304 podStartE2EDuration="3.569156304s" podCreationTimestamp="2026-02-20 08:21:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:23.566068538 +0000 UTC m=+932.540563358" watchObservedRunningTime="2026-02-20 08:21:23.569156304 +0000 UTC m=+932.543651134" Feb 20 08:21:23 crc kubenswrapper[4948]: I0220 08:21:23.596041 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" podStartSLOduration=3.5960244489999997 podStartE2EDuration="3.596024449s" podCreationTimestamp="2026-02-20 08:21:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:23.595136177 +0000 UTC m=+932.569631017" watchObservedRunningTime="2026-02-20 08:21:23.596024449 +0000 UTC m=+932.570519269" Feb 20 08:21:23 crc kubenswrapper[4948]: I0220 08:21:23.735682 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" path="/var/lib/kubelet/pods/5686ae39-4e44-4303-9b44-ad4b9ceae1ba/volumes" Feb 20 08:21:23 crc kubenswrapper[4948]: I0220 08:21:23.736385 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="977992e9-30ca-4420-98ae-92f91545cd10" path="/var/lib/kubelet/pods/977992e9-30ca-4420-98ae-92f91545cd10/volumes" Feb 20 08:21:24 crc kubenswrapper[4948]: I0220 08:21:24.557819 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"eeb3782c-0a88-4d17-b481-e2aef7ea386a","Type":"ContainerStarted","Data":"de9a37bb635341d457c0f9567c4af94f2752b0f6fb2a1c5b47992089cf013a01"} Feb 20 08:21:24 crc kubenswrapper[4948]: I0220 08:21:24.558095 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"eeb3782c-0a88-4d17-b481-e2aef7ea386a","Type":"ContainerStarted","Data":"ad5f6f0c61518eb7ded2c63c3aa9c6b2d5905b20178bc434822447b5304588eb"} Feb 20 08:21:24 crc kubenswrapper[4948]: I0220 08:21:24.558318 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:24 crc kubenswrapper[4948]: I0220 08:21:24.589662 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.128991371 podStartE2EDuration="4.589644578s" podCreationTimestamp="2026-02-20 08:21:20 +0000 UTC" firstStartedPulling="2026-02-20 08:21:21.87829986 +0000 UTC m=+930.852794680" lastFinishedPulling="2026-02-20 08:21:23.338953067 +0000 UTC m=+932.313447887" observedRunningTime="2026-02-20 08:21:24.582508822 +0000 UTC m=+933.557003642" watchObservedRunningTime="2026-02-20 08:21:24.589644578 +0000 UTC m=+933.564139398" Feb 20 08:21:25 crc kubenswrapper[4948]: I0220 08:21:25.519601 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 20 08:21:25 crc kubenswrapper[4948]: I0220 08:21:25.519667 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 20 08:21:25 crc kubenswrapper[4948]: I0220 08:21:25.567486 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 20 08:21:25 crc kubenswrapper[4948]: I0220 08:21:25.634923 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 20 08:21:25 crc kubenswrapper[4948]: I0220 08:21:25.720462 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 20 08:21:25 crc kubenswrapper[4948]: I0220 08:21:25.853723 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:21:25 crc kubenswrapper[4948]: I0220 08:21:25.854156 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:21:25 crc kubenswrapper[4948]: I0220 08:21:25.920903 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:21:26 crc kubenswrapper[4948]: I0220 08:21:26.617742 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:21:26 crc kubenswrapper[4948]: I0220 08:21:26.683299 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rt8tp"] Feb 20 08:21:26 crc kubenswrapper[4948]: I0220 08:21:26.898074 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 20 08:21:26 crc kubenswrapper[4948]: I0220 08:21:26.898407 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.106189 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-eb36-account-create-update-554g5"] Feb 20 08:21:28 crc kubenswrapper[4948]: E0220 08:21:28.106721 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="977992e9-30ca-4420-98ae-92f91545cd10" containerName="init" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.106733 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="977992e9-30ca-4420-98ae-92f91545cd10" containerName="init" Feb 20 08:21:28 crc kubenswrapper[4948]: E0220 08:21:28.106752 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" containerName="init" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.106758 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" containerName="init" Feb 20 08:21:28 crc kubenswrapper[4948]: E0220 08:21:28.106773 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="977992e9-30ca-4420-98ae-92f91545cd10" containerName="dnsmasq-dns" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.106778 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="977992e9-30ca-4420-98ae-92f91545cd10" containerName="dnsmasq-dns" Feb 20 08:21:28 crc kubenswrapper[4948]: E0220 08:21:28.106798 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" containerName="dnsmasq-dns" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.106803 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" containerName="dnsmasq-dns" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.106955 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="5686ae39-4e44-4303-9b44-ad4b9ceae1ba" containerName="dnsmasq-dns" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.106987 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="977992e9-30ca-4420-98ae-92f91545cd10" containerName="dnsmasq-dns" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.107422 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.109040 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.115191 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-eb36-account-create-update-554g5"] Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.174666 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-gm6g6"] Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.175728 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.184786 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gm6g6"] Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.253653 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-operator-scripts\") pod \"keystone-eb36-account-create-update-554g5\" (UID: \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\") " pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.253729 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4wx7\" (UniqueName: \"kubernetes.io/projected/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-kube-api-access-j4wx7\") pod \"keystone-eb36-account-create-update-554g5\" (UID: \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\") " pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.278208 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-bv65d"] Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.279168 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-bv65d" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.289501 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-bv65d"] Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.302183 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.355659 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89k8g\" (UniqueName: \"kubernetes.io/projected/52238644-5964-427e-8b77-9ca228d031db-kube-api-access-89k8g\") pod \"keystone-db-create-gm6g6\" (UID: \"52238644-5964-427e-8b77-9ca228d031db\") " pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.355793 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52238644-5964-427e-8b77-9ca228d031db-operator-scripts\") pod \"keystone-db-create-gm6g6\" (UID: \"52238644-5964-427e-8b77-9ca228d031db\") " pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.355984 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-operator-scripts\") pod \"keystone-eb36-account-create-update-554g5\" (UID: \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\") " pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.356094 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4wx7\" (UniqueName: \"kubernetes.io/projected/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-kube-api-access-j4wx7\") pod \"keystone-eb36-account-create-update-554g5\" (UID: \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\") " pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.357281 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-operator-scripts\") pod \"keystone-eb36-account-create-update-554g5\" (UID: \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\") " pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.383395 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6d58-account-create-update-bbx44"] Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.384609 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4wx7\" (UniqueName: \"kubernetes.io/projected/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-kube-api-access-j4wx7\") pod \"keystone-eb36-account-create-update-554g5\" (UID: \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\") " pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.384808 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.393184 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.410424 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6d58-account-create-update-bbx44"] Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.425042 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.457720 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89k8g\" (UniqueName: \"kubernetes.io/projected/52238644-5964-427e-8b77-9ca228d031db-kube-api-access-89k8g\") pod \"keystone-db-create-gm6g6\" (UID: \"52238644-5964-427e-8b77-9ca228d031db\") " pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.457793 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52238644-5964-427e-8b77-9ca228d031db-operator-scripts\") pod \"keystone-db-create-gm6g6\" (UID: \"52238644-5964-427e-8b77-9ca228d031db\") " pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.457920 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-operator-scripts\") pod \"placement-db-create-bv65d\" (UID: \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\") " pod="openstack/placement-db-create-bv65d" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.457983 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnscr\" (UniqueName: \"kubernetes.io/projected/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-kube-api-access-wnscr\") pod \"placement-db-create-bv65d\" (UID: \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\") " pod="openstack/placement-db-create-bv65d" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.458818 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52238644-5964-427e-8b77-9ca228d031db-operator-scripts\") pod \"keystone-db-create-gm6g6\" (UID: \"52238644-5964-427e-8b77-9ca228d031db\") " pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.473910 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89k8g\" (UniqueName: \"kubernetes.io/projected/52238644-5964-427e-8b77-9ca228d031db-kube-api-access-89k8g\") pod \"keystone-db-create-gm6g6\" (UID: \"52238644-5964-427e-8b77-9ca228d031db\") " pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.513907 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.560108 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnscr\" (UniqueName: \"kubernetes.io/projected/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-kube-api-access-wnscr\") pod \"placement-db-create-bv65d\" (UID: \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\") " pod="openstack/placement-db-create-bv65d" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.560442 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kqvl\" (UniqueName: \"kubernetes.io/projected/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-kube-api-access-2kqvl\") pod \"placement-6d58-account-create-update-bbx44\" (UID: \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\") " pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.560579 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-operator-scripts\") pod \"placement-6d58-account-create-update-bbx44\" (UID: \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\") " pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.561148 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-operator-scripts\") pod \"placement-db-create-bv65d\" (UID: \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\") " pod="openstack/placement-db-create-bv65d" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.561727 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-operator-scripts\") pod \"placement-db-create-bv65d\" (UID: \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\") " pod="openstack/placement-db-create-bv65d" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.595819 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnscr\" (UniqueName: \"kubernetes.io/projected/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-kube-api-access-wnscr\") pod \"placement-db-create-bv65d\" (UID: \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\") " pod="openstack/placement-db-create-bv65d" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.598034 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rt8tp" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerName="registry-server" containerID="cri-o://fbf3b4c07c31284ef058431c15f6fb31624a27a446740ecaf65fcf08b863a6e3" gracePeriod=2 Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.664103 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-operator-scripts\") pod \"placement-6d58-account-create-update-bbx44\" (UID: \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\") " pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.665652 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-operator-scripts\") pod \"placement-6d58-account-create-update-bbx44\" (UID: \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\") " pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.665846 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kqvl\" (UniqueName: \"kubernetes.io/projected/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-kube-api-access-2kqvl\") pod \"placement-6d58-account-create-update-bbx44\" (UID: \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\") " pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.687287 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kqvl\" (UniqueName: \"kubernetes.io/projected/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-kube-api-access-2kqvl\") pod \"placement-6d58-account-create-update-bbx44\" (UID: \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\") " pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.705247 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.722027 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.894340 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-bv65d" Feb 20 08:21:28 crc kubenswrapper[4948]: I0220 08:21:28.905789 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-eb36-account-create-update-554g5"] Feb 20 08:21:28 crc kubenswrapper[4948]: W0220 08:21:28.912934 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5bf7a21_af5d_4334_bc98_bd0f9eb9035a.slice/crio-25d97fb00e18b6ce84c602bbc6e333780c0537c4cab9cb2b3ce67336516b898c WatchSource:0}: Error finding container 25d97fb00e18b6ce84c602bbc6e333780c0537c4cab9cb2b3ce67336516b898c: Status 404 returned error can't find the container with id 25d97fb00e18b6ce84c602bbc6e333780c0537c4cab9cb2b3ce67336516b898c Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.038591 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gm6g6"] Feb 20 08:21:29 crc kubenswrapper[4948]: W0220 08:21:29.067824 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52238644_5964_427e_8b77_9ca228d031db.slice/crio-79852731770fecb861112fde754b01ca4c49d355a45bb283f9e41b055faf6747 WatchSource:0}: Error finding container 79852731770fecb861112fde754b01ca4c49d355a45bb283f9e41b055faf6747: Status 404 returned error can't find the container with id 79852731770fecb861112fde754b01ca4c49d355a45bb283f9e41b055faf6747 Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.173853 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6d58-account-create-update-bbx44"] Feb 20 08:21:29 crc kubenswrapper[4948]: W0220 08:21:29.181928 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1700d74_b484_4c6e_8fd3_d9dab8f4b7eb.slice/crio-fe0235453e94ed6b67ebdad926c4872aff88ca5acd235d564a2517d91ee0dafd WatchSource:0}: Error finding container fe0235453e94ed6b67ebdad926c4872aff88ca5acd235d564a2517d91ee0dafd: Status 404 returned error can't find the container with id fe0235453e94ed6b67ebdad926c4872aff88ca5acd235d564a2517d91ee0dafd Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.373185 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-bv65d"] Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.387251 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 20 08:21:29 crc kubenswrapper[4948]: W0220 08:21:29.396025 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf917fa7f_51f9_4c8f_bee2_11529a6d44a6.slice/crio-7f25c2e14bdc30bc83d90e8cf3dae79ddb8cda1ee4900ffa7eb5001c2eb0fd98 WatchSource:0}: Error finding container 7f25c2e14bdc30bc83d90e8cf3dae79ddb8cda1ee4900ffa7eb5001c2eb0fd98: Status 404 returned error can't find the container with id 7f25c2e14bdc30bc83d90e8cf3dae79ddb8cda1ee4900ffa7eb5001c2eb0fd98 Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.463706 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-mvb6f"] Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.463916 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" podUID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" containerName="dnsmasq-dns" containerID="cri-o://ae779bfaf190f36bd8dce0329a6a4c0ac105d70cb79287b3be92631d581020d5" gracePeriod=10 Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.465427 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.514041 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-tkqzb"] Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.515708 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.532520 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tkqzb"] Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.610332 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gm6g6" event={"ID":"52238644-5964-427e-8b77-9ca228d031db","Type":"ContainerStarted","Data":"79852731770fecb861112fde754b01ca4c49d355a45bb283f9e41b055faf6747"} Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.611161 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-eb36-account-create-update-554g5" event={"ID":"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a","Type":"ContainerStarted","Data":"25d97fb00e18b6ce84c602bbc6e333780c0537c4cab9cb2b3ce67336516b898c"} Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.612324 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6d58-account-create-update-bbx44" event={"ID":"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb","Type":"ContainerStarted","Data":"fe0235453e94ed6b67ebdad926c4872aff88ca5acd235d564a2517d91ee0dafd"} Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.613025 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-bv65d" event={"ID":"f917fa7f-51f9-4c8f-bee2-11529a6d44a6","Type":"ContainerStarted","Data":"7f25c2e14bdc30bc83d90e8cf3dae79ddb8cda1ee4900ffa7eb5001c2eb0fd98"} Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.615492 4948 generic.go:334] "Generic (PLEG): container finished" podID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerID="fbf3b4c07c31284ef058431c15f6fb31624a27a446740ecaf65fcf08b863a6e3" exitCode=0 Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.615566 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt8tp" event={"ID":"93f24735-d27d-4867-b9f7-cce47e7f376f","Type":"ContainerDied","Data":"fbf3b4c07c31284ef058431c15f6fb31624a27a446740ecaf65fcf08b863a6e3"} Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.632386 4948 generic.go:334] "Generic (PLEG): container finished" podID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" containerID="ae779bfaf190f36bd8dce0329a6a4c0ac105d70cb79287b3be92631d581020d5" exitCode=0 Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.633247 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" event={"ID":"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be","Type":"ContainerDied","Data":"ae779bfaf190f36bd8dce0329a6a4c0ac105d70cb79287b3be92631d581020d5"} Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.686792 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wc2m\" (UniqueName: \"kubernetes.io/projected/da816a64-160f-4a5c-b945-b8fa467208d6-kube-api-access-4wc2m\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.687177 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-config\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.687208 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-dns-svc\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.687235 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.687336 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.788225 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-config\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.788650 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-dns-svc\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.788676 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.789266 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-config\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.789376 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-dns-svc\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.789647 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.790079 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.790176 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wc2m\" (UniqueName: \"kubernetes.io/projected/da816a64-160f-4a5c-b945-b8fa467208d6-kube-api-access-4wc2m\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.790617 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.808625 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wc2m\" (UniqueName: \"kubernetes.io/projected/da816a64-160f-4a5c-b945-b8fa467208d6-kube-api-access-4wc2m\") pod \"dnsmasq-dns-698758b865-tkqzb\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:29 crc kubenswrapper[4948]: I0220 08:21:29.847300 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.274959 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tkqzb"] Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.427412 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.589560 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Feb 20 08:21:30 crc kubenswrapper[4948]: E0220 08:21:30.596366 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerName="extract-utilities" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.596399 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerName="extract-utilities" Feb 20 08:21:30 crc kubenswrapper[4948]: E0220 08:21:30.596413 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerName="registry-server" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.596419 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerName="registry-server" Feb 20 08:21:30 crc kubenswrapper[4948]: E0220 08:21:30.596429 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerName="extract-content" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.596435 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerName="extract-content" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.596579 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" containerName="registry-server" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.601196 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.607266 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.607391 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.607419 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-8lns4" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.607604 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.608909 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-catalog-content\") pod \"93f24735-d27d-4867-b9f7-cce47e7f376f\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.608965 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2cfc\" (UniqueName: \"kubernetes.io/projected/93f24735-d27d-4867-b9f7-cce47e7f376f-kube-api-access-q2cfc\") pod \"93f24735-d27d-4867-b9f7-cce47e7f376f\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.609081 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-utilities\") pod \"93f24735-d27d-4867-b9f7-cce47e7f376f\" (UID: \"93f24735-d27d-4867-b9f7-cce47e7f376f\") " Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.610148 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-utilities" (OuterVolumeSpecName: "utilities") pod "93f24735-d27d-4867-b9f7-cce47e7f376f" (UID: "93f24735-d27d-4867-b9f7-cce47e7f376f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.612607 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.619171 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93f24735-d27d-4867-b9f7-cce47e7f376f-kube-api-access-q2cfc" (OuterVolumeSpecName: "kube-api-access-q2cfc") pod "93f24735-d27d-4867-b9f7-cce47e7f376f" (UID: "93f24735-d27d-4867-b9f7-cce47e7f376f"). InnerVolumeSpecName "kube-api-access-q2cfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.648811 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-eb36-account-create-update-554g5" event={"ID":"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a","Type":"ContainerStarted","Data":"f4eb6ffc3a340bf3fb020244f7c41526c57d80da4f478c0885a3916041b2da30"} Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.660057 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6d58-account-create-update-bbx44" event={"ID":"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb","Type":"ContainerStarted","Data":"043fb8ebd535abc3ac64988e423e5b76f657b01785d57b024ae113afd44b4bf7"} Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.670809 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tkqzb" event={"ID":"da816a64-160f-4a5c-b945-b8fa467208d6","Type":"ContainerStarted","Data":"32b2c84d98711b304188499ff3c33cd5bf7af6d44c7a701d021f0522c169184b"} Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.680318 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-bv65d" event={"ID":"f917fa7f-51f9-4c8f-bee2-11529a6d44a6","Type":"ContainerStarted","Data":"279cb01f22f256d8185916f133561b9dbb69d1fb0ef3e531faed5ee14470b9d2"} Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.682624 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt8tp" event={"ID":"93f24735-d27d-4867-b9f7-cce47e7f376f","Type":"ContainerDied","Data":"104e51617975f8f0d6bf864e106e2486e0a18298cc9a24f078002083a6d423b2"} Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.682668 4948 scope.go:117] "RemoveContainer" containerID="fbf3b4c07c31284ef058431c15f6fb31624a27a446740ecaf65fcf08b863a6e3" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.682755 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rt8tp" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.687950 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gm6g6" event={"ID":"52238644-5964-427e-8b77-9ca228d031db","Type":"ContainerStarted","Data":"eb75184d7c11178b70689bbef778a20eca59576d9cf6e2826963a0598aa904a4"} Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.701068 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-eb36-account-create-update-554g5" podStartSLOduration=2.701052507 podStartE2EDuration="2.701052507s" podCreationTimestamp="2026-02-20 08:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:30.699405706 +0000 UTC m=+939.673900526" watchObservedRunningTime="2026-02-20 08:21:30.701052507 +0000 UTC m=+939.675547327" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.709339 4948 scope.go:117] "RemoveContainer" containerID="7f925a093fb8a5b68097ec9940401bb5483d01b61401a6a70fda1f29efe6615e" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.710229 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.710283 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4d60def-5dc9-460b-92ed-e3d0157982ec-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.710314 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e4d60def-5dc9-460b-92ed-e3d0157982ec-cache\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.710362 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.710384 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvm4h\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-kube-api-access-vvm4h\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.710403 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e4d60def-5dc9-460b-92ed-e3d0157982ec-lock\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.710439 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2cfc\" (UniqueName: \"kubernetes.io/projected/93f24735-d27d-4867-b9f7-cce47e7f376f-kube-api-access-q2cfc\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.710449 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.735463 4948 scope.go:117] "RemoveContainer" containerID="b689fa43e3b5c02fa357d00443dc58a7d6fa6d71f6692b435028801366e8a5e6" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.738413 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6d58-account-create-update-bbx44" podStartSLOduration=2.738398541 podStartE2EDuration="2.738398541s" podCreationTimestamp="2026-02-20 08:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:30.733635633 +0000 UTC m=+939.708130453" watchObservedRunningTime="2026-02-20 08:21:30.738398541 +0000 UTC m=+939.712893361" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.765879 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-bv65d" podStartSLOduration=2.76585781 podStartE2EDuration="2.76585781s" podCreationTimestamp="2026-02-20 08:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:30.755720539 +0000 UTC m=+939.730215359" watchObservedRunningTime="2026-02-20 08:21:30.76585781 +0000 UTC m=+939.740352630" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.793904 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-gm6g6" podStartSLOduration=2.793887644 podStartE2EDuration="2.793887644s" podCreationTimestamp="2026-02-20 08:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:30.780591955 +0000 UTC m=+939.755086775" watchObservedRunningTime="2026-02-20 08:21:30.793887644 +0000 UTC m=+939.768382464" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.812524 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.812591 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvm4h\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-kube-api-access-vvm4h\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.812621 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e4d60def-5dc9-460b-92ed-e3d0157982ec-lock\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.812671 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.812771 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4d60def-5dc9-460b-92ed-e3d0157982ec-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.812813 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e4d60def-5dc9-460b-92ed-e3d0157982ec-cache\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.813697 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: E0220 08:21:30.816528 4948 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 20 08:21:30 crc kubenswrapper[4948]: E0220 08:21:30.816552 4948 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 20 08:21:30 crc kubenswrapper[4948]: E0220 08:21:30.816600 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift podName:e4d60def-5dc9-460b-92ed-e3d0157982ec nodeName:}" failed. No retries permitted until 2026-02-20 08:21:31.316584146 +0000 UTC m=+940.291078966 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift") pod "swift-storage-0" (UID: "e4d60def-5dc9-460b-92ed-e3d0157982ec") : configmap "swift-ring-files" not found Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.816828 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e4d60def-5dc9-460b-92ed-e3d0157982ec-lock\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.826121 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e4d60def-5dc9-460b-92ed-e3d0157982ec-cache\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.826774 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4d60def-5dc9-460b-92ed-e3d0157982ec-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.832184 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvm4h\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-kube-api-access-vvm4h\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.839922 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.854347 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93f24735-d27d-4867-b9f7-cce47e7f376f" (UID: "93f24735-d27d-4867-b9f7-cce47e7f376f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:21:30 crc kubenswrapper[4948]: I0220 08:21:30.913904 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93f24735-d27d-4867-b9f7-cce47e7f376f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.043526 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-f44f4"] Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.044472 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.046043 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.046344 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.046534 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.088113 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-f44f4"] Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.101483 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-f44f4"] Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.108434 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-dllls"] Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.109708 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.117127 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-combined-ca-bundle\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.117202 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-swiftconf\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.117270 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-scripts\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.117381 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px2hh\" (UniqueName: \"kubernetes.io/projected/adf79be9-a443-4d51-b854-2067624fc60f-kube-api-access-px2hh\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.117518 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/adf79be9-a443-4d51-b854-2067624fc60f-etc-swift\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.117598 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-ring-data-devices\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.117660 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-dispersionconf\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.129083 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-dllls"] Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.137656 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:31 crc kubenswrapper[4948]: E0220 08:21:31.150645 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-px2hh ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-f44f4" podUID="adf79be9-a443-4d51-b854-2067624fc60f" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.162378 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rt8tp"] Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.168656 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rt8tp"] Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.219566 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-dns-svc\") pod \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.219717 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-ovsdbserver-nb\") pod \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.219738 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config\") pod \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.219772 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cpkv\" (UniqueName: \"kubernetes.io/projected/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-kube-api-access-9cpkv\") pod \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220006 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-scripts\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220036 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-swiftconf\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220084 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-dispersionconf\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220104 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-scripts\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220120 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px2hh\" (UniqueName: \"kubernetes.io/projected/adf79be9-a443-4d51-b854-2067624fc60f-kube-api-access-px2hh\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220143 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-ring-data-devices\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220178 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/adf79be9-a443-4d51-b854-2067624fc60f-etc-swift\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220199 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-ring-data-devices\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220215 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-dispersionconf\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220228 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-etc-swift\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220256 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-combined-ca-bundle\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220272 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89jgn\" (UniqueName: \"kubernetes.io/projected/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-kube-api-access-89jgn\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220296 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-swiftconf\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.220331 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-combined-ca-bundle\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.221549 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/adf79be9-a443-4d51-b854-2067624fc60f-etc-swift\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.221747 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-ring-data-devices\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.222000 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-scripts\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.224097 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-dispersionconf\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.225314 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-swiftconf\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.226301 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-combined-ca-bundle\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.233758 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-kube-api-access-9cpkv" (OuterVolumeSpecName: "kube-api-access-9cpkv") pod "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" (UID: "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be"). InnerVolumeSpecName "kube-api-access-9cpkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.237485 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px2hh\" (UniqueName: \"kubernetes.io/projected/adf79be9-a443-4d51-b854-2067624fc60f-kube-api-access-px2hh\") pod \"swift-ring-rebalance-f44f4\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.258033 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" (UID: "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: E0220 08:21:31.263238 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config podName:8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be nodeName:}" failed. No retries permitted until 2026-02-20 08:21:31.763198628 +0000 UTC m=+940.737693448 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config") pod "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" (UID: "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be") : error deleting /var/lib/kubelet/pods/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be/volume-subpaths: remove /var/lib/kubelet/pods/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be/volume-subpaths: no such file or directory Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.263536 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" (UID: "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.322210 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-ring-data-devices\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.322256 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:31 crc kubenswrapper[4948]: E0220 08:21:31.322376 4948 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 20 08:21:31 crc kubenswrapper[4948]: E0220 08:21:31.322396 4948 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 20 08:21:31 crc kubenswrapper[4948]: E0220 08:21:31.322439 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift podName:e4d60def-5dc9-460b-92ed-e3d0157982ec nodeName:}" failed. No retries permitted until 2026-02-20 08:21:32.322424044 +0000 UTC m=+941.296918864 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift") pod "swift-storage-0" (UID: "e4d60def-5dc9-460b-92ed-e3d0157982ec") : configmap "swift-ring-files" not found Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.322795 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-etc-swift\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.322906 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-ring-data-devices\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.323056 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-etc-swift\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.323110 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-combined-ca-bundle\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.323129 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89jgn\" (UniqueName: \"kubernetes.io/projected/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-kube-api-access-89jgn\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.323691 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-swiftconf\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.323783 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-scripts\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.323845 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-dispersionconf\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.323988 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.324000 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.324011 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cpkv\" (UniqueName: \"kubernetes.io/projected/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-kube-api-access-9cpkv\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.324279 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-scripts\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.333410 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-dispersionconf\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.333549 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-swiftconf\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.333734 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-combined-ca-bundle\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.346101 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89jgn\" (UniqueName: \"kubernetes.io/projected/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-kube-api-access-89jgn\") pod \"swift-ring-rebalance-dllls\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.376184 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.463123 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.712166 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.712307 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-mvb6f" event={"ID":"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be","Type":"ContainerDied","Data":"a0b4a492fe4751057387e9dea455a523858f53f3d884620fffd805ea3be59755"} Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.712653 4948 scope.go:117] "RemoveContainer" containerID="ae779bfaf190f36bd8dce0329a6a4c0ac105d70cb79287b3be92631d581020d5" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.714015 4948 generic.go:334] "Generic (PLEG): container finished" podID="52238644-5964-427e-8b77-9ca228d031db" containerID="eb75184d7c11178b70689bbef778a20eca59576d9cf6e2826963a0598aa904a4" exitCode=0 Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.714155 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gm6g6" event={"ID":"52238644-5964-427e-8b77-9ca228d031db","Type":"ContainerDied","Data":"eb75184d7c11178b70689bbef778a20eca59576d9cf6e2826963a0598aa904a4"} Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.716626 4948 generic.go:334] "Generic (PLEG): container finished" podID="b5bf7a21-af5d-4334-bc98-bd0f9eb9035a" containerID="f4eb6ffc3a340bf3fb020244f7c41526c57d80da4f478c0885a3916041b2da30" exitCode=0 Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.716749 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-eb36-account-create-update-554g5" event={"ID":"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a","Type":"ContainerDied","Data":"f4eb6ffc3a340bf3fb020244f7c41526c57d80da4f478c0885a3916041b2da30"} Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.718301 4948 generic.go:334] "Generic (PLEG): container finished" podID="e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb" containerID="043fb8ebd535abc3ac64988e423e5b76f657b01785d57b024ae113afd44b4bf7" exitCode=0 Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.718346 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6d58-account-create-update-bbx44" event={"ID":"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb","Type":"ContainerDied","Data":"043fb8ebd535abc3ac64988e423e5b76f657b01785d57b024ae113afd44b4bf7"} Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.720119 4948 generic.go:334] "Generic (PLEG): container finished" podID="da816a64-160f-4a5c-b945-b8fa467208d6" containerID="48592a00a61a9ce774a176445c59d2e205743cbfe0a6d5b932c49a8ff31d312a" exitCode=0 Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.720197 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tkqzb" event={"ID":"da816a64-160f-4a5c-b945-b8fa467208d6","Type":"ContainerDied","Data":"48592a00a61a9ce774a176445c59d2e205743cbfe0a6d5b932c49a8ff31d312a"} Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.721784 4948 generic.go:334] "Generic (PLEG): container finished" podID="f917fa7f-51f9-4c8f-bee2-11529a6d44a6" containerID="279cb01f22f256d8185916f133561b9dbb69d1fb0ef3e531faed5ee14470b9d2" exitCode=0 Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.721931 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.736403 4948 scope.go:117] "RemoveContainer" containerID="c624ced0b3cd9432f64567b57ff22f472c7a407536a9a515e94a3ab95cf57662" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.745178 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93f24735-d27d-4867-b9f7-cce47e7f376f" path="/var/lib/kubelet/pods/93f24735-d27d-4867-b9f7-cce47e7f376f/volumes" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.746839 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.829083 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-bv65d" event={"ID":"f917fa7f-51f9-4c8f-bee2-11529a6d44a6","Type":"ContainerDied","Data":"279cb01f22f256d8185916f133561b9dbb69d1fb0ef3e531faed5ee14470b9d2"} Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.837881 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/adf79be9-a443-4d51-b854-2067624fc60f-etc-swift\") pod \"adf79be9-a443-4d51-b854-2067624fc60f\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.837930 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-ring-data-devices\") pod \"adf79be9-a443-4d51-b854-2067624fc60f\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.838042 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px2hh\" (UniqueName: \"kubernetes.io/projected/adf79be9-a443-4d51-b854-2067624fc60f-kube-api-access-px2hh\") pod \"adf79be9-a443-4d51-b854-2067624fc60f\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.838074 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-combined-ca-bundle\") pod \"adf79be9-a443-4d51-b854-2067624fc60f\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.838130 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-swiftconf\") pod \"adf79be9-a443-4d51-b854-2067624fc60f\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.838166 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-scripts\") pod \"adf79be9-a443-4d51-b854-2067624fc60f\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.838229 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config\") pod \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\" (UID: \"8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.838250 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-dispersionconf\") pod \"adf79be9-a443-4d51-b854-2067624fc60f\" (UID: \"adf79be9-a443-4d51-b854-2067624fc60f\") " Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.838949 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "adf79be9-a443-4d51-b854-2067624fc60f" (UID: "adf79be9-a443-4d51-b854-2067624fc60f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.839256 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adf79be9-a443-4d51-b854-2067624fc60f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "adf79be9-a443-4d51-b854-2067624fc60f" (UID: "adf79be9-a443-4d51-b854-2067624fc60f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.839910 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config" (OuterVolumeSpecName: "config") pod "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" (UID: "8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.840909 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-scripts" (OuterVolumeSpecName: "scripts") pod "adf79be9-a443-4d51-b854-2067624fc60f" (UID: "adf79be9-a443-4d51-b854-2067624fc60f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.842320 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "adf79be9-a443-4d51-b854-2067624fc60f" (UID: "adf79be9-a443-4d51-b854-2067624fc60f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.843062 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "adf79be9-a443-4d51-b854-2067624fc60f" (UID: "adf79be9-a443-4d51-b854-2067624fc60f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.843945 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "adf79be9-a443-4d51-b854-2067624fc60f" (UID: "adf79be9-a443-4d51-b854-2067624fc60f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.845088 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adf79be9-a443-4d51-b854-2067624fc60f-kube-api-access-px2hh" (OuterVolumeSpecName: "kube-api-access-px2hh") pod "adf79be9-a443-4d51-b854-2067624fc60f" (UID: "adf79be9-a443-4d51-b854-2067624fc60f"). InnerVolumeSpecName "kube-api-access-px2hh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.903877 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-dllls"] Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.940923 4948 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/adf79be9-a443-4d51-b854-2067624fc60f-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.940962 4948 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.940991 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px2hh\" (UniqueName: \"kubernetes.io/projected/adf79be9-a443-4d51-b854-2067624fc60f-kube-api-access-px2hh\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.941003 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.941014 4948 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.941027 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/adf79be9-a443-4d51-b854-2067624fc60f-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.941038 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:31 crc kubenswrapper[4948]: I0220 08:21:31.941049 4948 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/adf79be9-a443-4d51-b854-2067624fc60f-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.050380 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-mvb6f"] Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.058408 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-mvb6f"] Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.319821 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-fs9rr"] Feb 20 08:21:32 crc kubenswrapper[4948]: E0220 08:21:32.320276 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" containerName="init" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.320303 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" containerName="init" Feb 20 08:21:32 crc kubenswrapper[4948]: E0220 08:21:32.320327 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" containerName="dnsmasq-dns" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.320338 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" containerName="dnsmasq-dns" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.320586 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" containerName="dnsmasq-dns" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.321324 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.341550 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-fs9rr"] Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.349510 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:32 crc kubenswrapper[4948]: E0220 08:21:32.349783 4948 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 20 08:21:32 crc kubenswrapper[4948]: E0220 08:21:32.349816 4948 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 20 08:21:32 crc kubenswrapper[4948]: E0220 08:21:32.349887 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift podName:e4d60def-5dc9-460b-92ed-e3d0157982ec nodeName:}" failed. No retries permitted until 2026-02-20 08:21:34.349860869 +0000 UTC m=+943.324355719 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift") pod "swift-storage-0" (UID: "e4d60def-5dc9-460b-92ed-e3d0157982ec") : configmap "swift-ring-files" not found Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.422790 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-c08d-account-create-update-54vb7"] Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.424359 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.431327 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.443704 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c08d-account-create-update-54vb7"] Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.450604 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb6wl\" (UniqueName: \"kubernetes.io/projected/4d815209-521c-4c7f-a026-18899832459f-kube-api-access-sb6wl\") pod \"glance-db-create-fs9rr\" (UID: \"4d815209-521c-4c7f-a026-18899832459f\") " pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.450828 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d815209-521c-4c7f-a026-18899832459f-operator-scripts\") pod \"glance-db-create-fs9rr\" (UID: \"4d815209-521c-4c7f-a026-18899832459f\") " pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.553943 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d815209-521c-4c7f-a026-18899832459f-operator-scripts\") pod \"glance-db-create-fs9rr\" (UID: \"4d815209-521c-4c7f-a026-18899832459f\") " pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.554111 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6bll\" (UniqueName: \"kubernetes.io/projected/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-kube-api-access-v6bll\") pod \"glance-c08d-account-create-update-54vb7\" (UID: \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\") " pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.554185 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb6wl\" (UniqueName: \"kubernetes.io/projected/4d815209-521c-4c7f-a026-18899832459f-kube-api-access-sb6wl\") pod \"glance-db-create-fs9rr\" (UID: \"4d815209-521c-4c7f-a026-18899832459f\") " pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.554231 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-operator-scripts\") pod \"glance-c08d-account-create-update-54vb7\" (UID: \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\") " pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.554944 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d815209-521c-4c7f-a026-18899832459f-operator-scripts\") pod \"glance-db-create-fs9rr\" (UID: \"4d815209-521c-4c7f-a026-18899832459f\") " pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.577863 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb6wl\" (UniqueName: \"kubernetes.io/projected/4d815209-521c-4c7f-a026-18899832459f-kube-api-access-sb6wl\") pod \"glance-db-create-fs9rr\" (UID: \"4d815209-521c-4c7f-a026-18899832459f\") " pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.637625 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.656945 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6bll\" (UniqueName: \"kubernetes.io/projected/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-kube-api-access-v6bll\") pod \"glance-c08d-account-create-update-54vb7\" (UID: \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\") " pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.657096 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-operator-scripts\") pod \"glance-c08d-account-create-update-54vb7\" (UID: \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\") " pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.657872 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-operator-scripts\") pod \"glance-c08d-account-create-update-54vb7\" (UID: \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\") " pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.680592 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6bll\" (UniqueName: \"kubernetes.io/projected/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-kube-api-access-v6bll\") pod \"glance-c08d-account-create-update-54vb7\" (UID: \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\") " pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.733034 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tkqzb" event={"ID":"da816a64-160f-4a5c-b945-b8fa467208d6","Type":"ContainerStarted","Data":"562bf68215847fe83f99073d009b1ee89904610f3d702b2ab282baecbca78fb2"} Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.734037 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.736960 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-dllls" event={"ID":"642adfd1-e253-4e0a-80e6-860a9f5c1ae1","Type":"ContainerStarted","Data":"befd58899320f920bf3b69e7a9c823d94bc5bd3a3193731f420fe312a97c42e6"} Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.738409 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-f44f4" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.767642 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-tkqzb" podStartSLOduration=3.767603417 podStartE2EDuration="3.767603417s" podCreationTimestamp="2026-02-20 08:21:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:32.766242953 +0000 UTC m=+941.740737773" watchObservedRunningTime="2026-02-20 08:21:32.767603417 +0000 UTC m=+941.742098237" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.795091 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.813504 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-f44f4"] Feb 20 08:21:32 crc kubenswrapper[4948]: I0220 08:21:32.828946 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-f44f4"] Feb 20 08:21:33 crc kubenswrapper[4948]: I0220 08:21:33.413871 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c08d-account-create-update-54vb7"] Feb 20 08:21:33 crc kubenswrapper[4948]: I0220 08:21:33.528209 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-fs9rr"] Feb 20 08:21:33 crc kubenswrapper[4948]: I0220 08:21:33.744916 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be" path="/var/lib/kubelet/pods/8e1cfe6e-46bb-457f-9a82-7e2a6ce4e3be/volumes" Feb 20 08:21:33 crc kubenswrapper[4948]: I0220 08:21:33.745633 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adf79be9-a443-4d51-b854-2067624fc60f" path="/var/lib/kubelet/pods/adf79be9-a443-4d51-b854-2067624fc60f/volumes" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.133044 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-rnvbh"] Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.134477 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.136450 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.152586 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-rnvbh"] Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.296819 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92714d74-2a44-42e9-8829-d99ac773ac75-operator-scripts\") pod \"root-account-create-update-rnvbh\" (UID: \"92714d74-2a44-42e9-8829-d99ac773ac75\") " pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.296964 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f22vf\" (UniqueName: \"kubernetes.io/projected/92714d74-2a44-42e9-8829-d99ac773ac75-kube-api-access-f22vf\") pod \"root-account-create-update-rnvbh\" (UID: \"92714d74-2a44-42e9-8829-d99ac773ac75\") " pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.398446 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f22vf\" (UniqueName: \"kubernetes.io/projected/92714d74-2a44-42e9-8829-d99ac773ac75-kube-api-access-f22vf\") pod \"root-account-create-update-rnvbh\" (UID: \"92714d74-2a44-42e9-8829-d99ac773ac75\") " pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.398548 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.398571 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92714d74-2a44-42e9-8829-d99ac773ac75-operator-scripts\") pod \"root-account-create-update-rnvbh\" (UID: \"92714d74-2a44-42e9-8829-d99ac773ac75\") " pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:34 crc kubenswrapper[4948]: E0220 08:21:34.398961 4948 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 20 08:21:34 crc kubenswrapper[4948]: E0220 08:21:34.399078 4948 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 20 08:21:34 crc kubenswrapper[4948]: E0220 08:21:34.399166 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift podName:e4d60def-5dc9-460b-92ed-e3d0157982ec nodeName:}" failed. No retries permitted until 2026-02-20 08:21:38.399136453 +0000 UTC m=+947.373631313 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift") pod "swift-storage-0" (UID: "e4d60def-5dc9-460b-92ed-e3d0157982ec") : configmap "swift-ring-files" not found Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.399252 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92714d74-2a44-42e9-8829-d99ac773ac75-operator-scripts\") pod \"root-account-create-update-rnvbh\" (UID: \"92714d74-2a44-42e9-8829-d99ac773ac75\") " pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.429059 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f22vf\" (UniqueName: \"kubernetes.io/projected/92714d74-2a44-42e9-8829-d99ac773ac75-kube-api-access-f22vf\") pod \"root-account-create-update-rnvbh\" (UID: \"92714d74-2a44-42e9-8829-d99ac773ac75\") " pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:34 crc kubenswrapper[4948]: I0220 08:21:34.465180 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:35 crc kubenswrapper[4948]: W0220 08:21:35.268862 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d815209_521c_4c7f_a026_18899832459f.slice/crio-83bd7271243acafa6581de499ba5b4200de0d064e497f5bdb3745d092e153305 WatchSource:0}: Error finding container 83bd7271243acafa6581de499ba5b4200de0d064e497f5bdb3745d092e153305: Status 404 returned error can't find the container with id 83bd7271243acafa6581de499ba5b4200de0d064e497f5bdb3745d092e153305 Feb 20 08:21:35 crc kubenswrapper[4948]: W0220 08:21:35.272172 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod333ba21a_82a9_4f39_9ec5_afb7ae2ccf60.slice/crio-7eca72d6b91c88e5a2e15a5744106cd7576ef511dae119df0c47c0377f312a2d WatchSource:0}: Error finding container 7eca72d6b91c88e5a2e15a5744106cd7576ef511dae119df0c47c0377f312a2d: Status 404 returned error can't find the container with id 7eca72d6b91c88e5a2e15a5744106cd7576ef511dae119df0c47c0377f312a2d Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.486269 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.531149 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-bv65d" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.562330 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.571278 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.619263 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnscr\" (UniqueName: \"kubernetes.io/projected/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-kube-api-access-wnscr\") pod \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\" (UID: \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\") " Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.619337 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-operator-scripts\") pod \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\" (UID: \"f917fa7f-51f9-4c8f-bee2-11529a6d44a6\") " Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.619392 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-operator-scripts\") pod \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\" (UID: \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\") " Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.619437 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4wx7\" (UniqueName: \"kubernetes.io/projected/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-kube-api-access-j4wx7\") pod \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\" (UID: \"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a\") " Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.620169 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f917fa7f-51f9-4c8f-bee2-11529a6d44a6" (UID: "f917fa7f-51f9-4c8f-bee2-11529a6d44a6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.620180 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b5bf7a21-af5d-4334-bc98-bd0f9eb9035a" (UID: "b5bf7a21-af5d-4334-bc98-bd0f9eb9035a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.620558 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.620577 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.625684 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-kube-api-access-wnscr" (OuterVolumeSpecName: "kube-api-access-wnscr") pod "f917fa7f-51f9-4c8f-bee2-11529a6d44a6" (UID: "f917fa7f-51f9-4c8f-bee2-11529a6d44a6"). InnerVolumeSpecName "kube-api-access-wnscr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.638662 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-kube-api-access-j4wx7" (OuterVolumeSpecName: "kube-api-access-j4wx7") pod "b5bf7a21-af5d-4334-bc98-bd0f9eb9035a" (UID: "b5bf7a21-af5d-4334-bc98-bd0f9eb9035a"). InnerVolumeSpecName "kube-api-access-j4wx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.721594 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kqvl\" (UniqueName: \"kubernetes.io/projected/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-kube-api-access-2kqvl\") pod \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\" (UID: \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\") " Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.721670 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52238644-5964-427e-8b77-9ca228d031db-operator-scripts\") pod \"52238644-5964-427e-8b77-9ca228d031db\" (UID: \"52238644-5964-427e-8b77-9ca228d031db\") " Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.721789 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89k8g\" (UniqueName: \"kubernetes.io/projected/52238644-5964-427e-8b77-9ca228d031db-kube-api-access-89k8g\") pod \"52238644-5964-427e-8b77-9ca228d031db\" (UID: \"52238644-5964-427e-8b77-9ca228d031db\") " Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.721826 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-operator-scripts\") pod \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\" (UID: \"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb\") " Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.722290 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnscr\" (UniqueName: \"kubernetes.io/projected/f917fa7f-51f9-4c8f-bee2-11529a6d44a6-kube-api-access-wnscr\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.722310 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4wx7\" (UniqueName: \"kubernetes.io/projected/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a-kube-api-access-j4wx7\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.722615 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb" (UID: "e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.723246 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52238644-5964-427e-8b77-9ca228d031db-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "52238644-5964-427e-8b77-9ca228d031db" (UID: "52238644-5964-427e-8b77-9ca228d031db"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.733790 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-kube-api-access-2kqvl" (OuterVolumeSpecName: "kube-api-access-2kqvl") pod "e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb" (UID: "e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb"). InnerVolumeSpecName "kube-api-access-2kqvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.734170 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52238644-5964-427e-8b77-9ca228d031db-kube-api-access-89k8g" (OuterVolumeSpecName: "kube-api-access-89k8g") pod "52238644-5964-427e-8b77-9ca228d031db" (UID: "52238644-5964-427e-8b77-9ca228d031db"). InnerVolumeSpecName "kube-api-access-89k8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.770308 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-rnvbh"] Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.774110 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fs9rr" event={"ID":"4d815209-521c-4c7f-a026-18899832459f","Type":"ContainerStarted","Data":"8e0c0f098f3973e04052c6dc5c4da1ff436370579d9b1ea6857d9e0c6bed683e"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.774153 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fs9rr" event={"ID":"4d815209-521c-4c7f-a026-18899832459f","Type":"ContainerStarted","Data":"83bd7271243acafa6581de499ba5b4200de0d064e497f5bdb3745d092e153305"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.788377 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gm6g6" event={"ID":"52238644-5964-427e-8b77-9ca228d031db","Type":"ContainerDied","Data":"79852731770fecb861112fde754b01ca4c49d355a45bb283f9e41b055faf6747"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.789625 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79852731770fecb861112fde754b01ca4c49d355a45bb283f9e41b055faf6747" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.788499 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gm6g6" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.795131 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-eb36-account-create-update-554g5" event={"ID":"b5bf7a21-af5d-4334-bc98-bd0f9eb9035a","Type":"ContainerDied","Data":"25d97fb00e18b6ce84c602bbc6e333780c0537c4cab9cb2b3ce67336516b898c"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.795178 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25d97fb00e18b6ce84c602bbc6e333780c0537c4cab9cb2b3ce67336516b898c" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.795417 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-eb36-account-create-update-554g5" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.798595 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6d58-account-create-update-bbx44" event={"ID":"e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb","Type":"ContainerDied","Data":"fe0235453e94ed6b67ebdad926c4872aff88ca5acd235d564a2517d91ee0dafd"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.798628 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe0235453e94ed6b67ebdad926c4872aff88ca5acd235d564a2517d91ee0dafd" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.798707 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6d58-account-create-update-bbx44" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.801370 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c08d-account-create-update-54vb7" event={"ID":"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60","Type":"ContainerStarted","Data":"67a7241e5369f13541e0b84997b90ee0fd96a87f376c9c419e0e637fa48068da"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.801441 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c08d-account-create-update-54vb7" event={"ID":"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60","Type":"ContainerStarted","Data":"7eca72d6b91c88e5a2e15a5744106cd7576ef511dae119df0c47c0377f312a2d"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.805498 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-dllls" event={"ID":"642adfd1-e253-4e0a-80e6-860a9f5c1ae1","Type":"ContainerStarted","Data":"8fbeb28cef63fcfd89707879381c2e4ed399e4426a2e6226acf2ce1d9b7e8e2c"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.807132 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-fs9rr" podStartSLOduration=3.807113914 podStartE2EDuration="3.807113914s" podCreationTimestamp="2026-02-20 08:21:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:35.790878413 +0000 UTC m=+944.765373253" watchObservedRunningTime="2026-02-20 08:21:35.807113914 +0000 UTC m=+944.781608734" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.808223 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-bv65d" event={"ID":"f917fa7f-51f9-4c8f-bee2-11529a6d44a6","Type":"ContainerDied","Data":"7f25c2e14bdc30bc83d90e8cf3dae79ddb8cda1ee4900ffa7eb5001c2eb0fd98"} Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.808253 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f25c2e14bdc30bc83d90e8cf3dae79ddb8cda1ee4900ffa7eb5001c2eb0fd98" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.808453 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-bv65d" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.820934 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-c08d-account-create-update-54vb7" podStartSLOduration=3.8209183859999998 podStartE2EDuration="3.820918386s" podCreationTimestamp="2026-02-20 08:21:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:35.820175058 +0000 UTC m=+944.794669878" watchObservedRunningTime="2026-02-20 08:21:35.820918386 +0000 UTC m=+944.795413206" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.823534 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kqvl\" (UniqueName: \"kubernetes.io/projected/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-kube-api-access-2kqvl\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.823557 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52238644-5964-427e-8b77-9ca228d031db-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.823567 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89k8g\" (UniqueName: \"kubernetes.io/projected/52238644-5964-427e-8b77-9ca228d031db-kube-api-access-89k8g\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.823576 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:35 crc kubenswrapper[4948]: I0220 08:21:35.840933 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-dllls" podStartSLOduration=1.426575126 podStartE2EDuration="4.840917511s" podCreationTimestamp="2026-02-20 08:21:31 +0000 UTC" firstStartedPulling="2026-02-20 08:21:31.908838665 +0000 UTC m=+940.883333495" lastFinishedPulling="2026-02-20 08:21:35.32318106 +0000 UTC m=+944.297675880" observedRunningTime="2026-02-20 08:21:35.83885855 +0000 UTC m=+944.813353360" watchObservedRunningTime="2026-02-20 08:21:35.840917511 +0000 UTC m=+944.815412331" Feb 20 08:21:36 crc kubenswrapper[4948]: E0220 08:21:36.382294 4948 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73b84bb7_f594_4823_ac03_40fdac6ee177.slice/crio-conmon-d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92714d74_2a44_42e9_8829_d99ac773ac75.slice/crio-5dfd385e001fae733c595da928ba9b46f6e3cc2e4d358e37840e650d21ce855a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73b84bb7_f594_4823_ac03_40fdac6ee177.slice/crio-d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364.scope\": RecentStats: unable to find data in memory cache]" Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.820426 4948 generic.go:334] "Generic (PLEG): container finished" podID="4d815209-521c-4c7f-a026-18899832459f" containerID="8e0c0f098f3973e04052c6dc5c4da1ff436370579d9b1ea6857d9e0c6bed683e" exitCode=0 Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.820522 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fs9rr" event={"ID":"4d815209-521c-4c7f-a026-18899832459f","Type":"ContainerDied","Data":"8e0c0f098f3973e04052c6dc5c4da1ff436370579d9b1ea6857d9e0c6bed683e"} Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.824164 4948 generic.go:334] "Generic (PLEG): container finished" podID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerID="d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364" exitCode=0 Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.824245 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"73b84bb7-f594-4823-ac03-40fdac6ee177","Type":"ContainerDied","Data":"d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364"} Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.829361 4948 generic.go:334] "Generic (PLEG): container finished" podID="92714d74-2a44-42e9-8829-d99ac773ac75" containerID="5dfd385e001fae733c595da928ba9b46f6e3cc2e4d358e37840e650d21ce855a" exitCode=0 Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.829453 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rnvbh" event={"ID":"92714d74-2a44-42e9-8829-d99ac773ac75","Type":"ContainerDied","Data":"5dfd385e001fae733c595da928ba9b46f6e3cc2e4d358e37840e650d21ce855a"} Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.829488 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rnvbh" event={"ID":"92714d74-2a44-42e9-8829-d99ac773ac75","Type":"ContainerStarted","Data":"5cee8fe5e9ab322142fa63fb214c8bbff5f35bb77dce1e1e2cae0a5fb25cc1c8"} Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.832106 4948 generic.go:334] "Generic (PLEG): container finished" podID="333ba21a-82a9-4f39-9ec5-afb7ae2ccf60" containerID="67a7241e5369f13541e0b84997b90ee0fd96a87f376c9c419e0e637fa48068da" exitCode=0 Feb 20 08:21:36 crc kubenswrapper[4948]: I0220 08:21:36.832621 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c08d-account-create-update-54vb7" event={"ID":"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60","Type":"ContainerDied","Data":"67a7241e5369f13541e0b84997b90ee0fd96a87f376c9c419e0e637fa48068da"} Feb 20 08:21:37 crc kubenswrapper[4948]: I0220 08:21:37.845491 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"73b84bb7-f594-4823-ac03-40fdac6ee177","Type":"ContainerStarted","Data":"f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85"} Feb 20 08:21:37 crc kubenswrapper[4948]: I0220 08:21:37.846049 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 20 08:21:37 crc kubenswrapper[4948]: I0220 08:21:37.848108 4948 generic.go:334] "Generic (PLEG): container finished" podID="790894c9-053c-497c-955d-ce7519111dd6" containerID="77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242" exitCode=0 Feb 20 08:21:37 crc kubenswrapper[4948]: I0220 08:21:37.848150 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"790894c9-053c-497c-955d-ce7519111dd6","Type":"ContainerDied","Data":"77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242"} Feb 20 08:21:37 crc kubenswrapper[4948]: I0220 08:21:37.877493 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.429138858 podStartE2EDuration="55.8774688s" podCreationTimestamp="2026-02-20 08:20:42 +0000 UTC" firstStartedPulling="2026-02-20 08:20:44.406206929 +0000 UTC m=+893.380701749" lastFinishedPulling="2026-02-20 08:21:02.854536861 +0000 UTC m=+911.829031691" observedRunningTime="2026-02-20 08:21:37.873259996 +0000 UTC m=+946.847754816" watchObservedRunningTime="2026-02-20 08:21:37.8774688 +0000 UTC m=+946.851963640" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.024852 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.026368 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.026605 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.027849 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"df8e056c06ac241c46a927c4e3f5a28f39d042bf13a8c672499d528786dbbbb0"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.027920 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://df8e056c06ac241c46a927c4e3f5a28f39d042bf13a8c672499d528786dbbbb0" gracePeriod=600 Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.373964 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.391079 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.423803 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.491058 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6wl\" (UniqueName: \"kubernetes.io/projected/4d815209-521c-4c7f-a026-18899832459f-kube-api-access-sb6wl\") pod \"4d815209-521c-4c7f-a026-18899832459f\" (UID: \"4d815209-521c-4c7f-a026-18899832459f\") " Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.491234 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-operator-scripts\") pod \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\" (UID: \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\") " Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.491355 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92714d74-2a44-42e9-8829-d99ac773ac75-operator-scripts\") pod \"92714d74-2a44-42e9-8829-d99ac773ac75\" (UID: \"92714d74-2a44-42e9-8829-d99ac773ac75\") " Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.491490 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d815209-521c-4c7f-a026-18899832459f-operator-scripts\") pod \"4d815209-521c-4c7f-a026-18899832459f\" (UID: \"4d815209-521c-4c7f-a026-18899832459f\") " Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.491515 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6bll\" (UniqueName: \"kubernetes.io/projected/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-kube-api-access-v6bll\") pod \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\" (UID: \"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60\") " Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.491603 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f22vf\" (UniqueName: \"kubernetes.io/projected/92714d74-2a44-42e9-8829-d99ac773ac75-kube-api-access-f22vf\") pod \"92714d74-2a44-42e9-8829-d99ac773ac75\" (UID: \"92714d74-2a44-42e9-8829-d99ac773ac75\") " Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.491831 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:38 crc kubenswrapper[4948]: E0220 08:21:38.492433 4948 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 20 08:21:38 crc kubenswrapper[4948]: E0220 08:21:38.492519 4948 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 20 08:21:38 crc kubenswrapper[4948]: E0220 08:21:38.492606 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift podName:e4d60def-5dc9-460b-92ed-e3d0157982ec nodeName:}" failed. No retries permitted until 2026-02-20 08:21:46.492591302 +0000 UTC m=+955.467086122 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift") pod "swift-storage-0" (UID: "e4d60def-5dc9-460b-92ed-e3d0157982ec") : configmap "swift-ring-files" not found Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.495410 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "333ba21a-82a9-4f39-9ec5-afb7ae2ccf60" (UID: "333ba21a-82a9-4f39-9ec5-afb7ae2ccf60"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.495480 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d815209-521c-4c7f-a026-18899832459f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4d815209-521c-4c7f-a026-18899832459f" (UID: "4d815209-521c-4c7f-a026-18899832459f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.495617 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92714d74-2a44-42e9-8829-d99ac773ac75-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "92714d74-2a44-42e9-8829-d99ac773ac75" (UID: "92714d74-2a44-42e9-8829-d99ac773ac75"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.501133 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d815209-521c-4c7f-a026-18899832459f-kube-api-access-sb6wl" (OuterVolumeSpecName: "kube-api-access-sb6wl") pod "4d815209-521c-4c7f-a026-18899832459f" (UID: "4d815209-521c-4c7f-a026-18899832459f"). InnerVolumeSpecName "kube-api-access-sb6wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.501170 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-kube-api-access-v6bll" (OuterVolumeSpecName: "kube-api-access-v6bll") pod "333ba21a-82a9-4f39-9ec5-afb7ae2ccf60" (UID: "333ba21a-82a9-4f39-9ec5-afb7ae2ccf60"). InnerVolumeSpecName "kube-api-access-v6bll". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.501187 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92714d74-2a44-42e9-8829-d99ac773ac75-kube-api-access-f22vf" (OuterVolumeSpecName: "kube-api-access-f22vf") pod "92714d74-2a44-42e9-8829-d99ac773ac75" (UID: "92714d74-2a44-42e9-8829-d99ac773ac75"). InnerVolumeSpecName "kube-api-access-f22vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.593843 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f22vf\" (UniqueName: \"kubernetes.io/projected/92714d74-2a44-42e9-8829-d99ac773ac75-kube-api-access-f22vf\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.594188 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6wl\" (UniqueName: \"kubernetes.io/projected/4d815209-521c-4c7f-a026-18899832459f-kube-api-access-sb6wl\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.594198 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.594207 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92714d74-2a44-42e9-8829-d99ac773ac75-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.594216 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d815209-521c-4c7f-a026-18899832459f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.594224 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6bll\" (UniqueName: \"kubernetes.io/projected/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60-kube-api-access-v6bll\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.857240 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="df8e056c06ac241c46a927c4e3f5a28f39d042bf13a8c672499d528786dbbbb0" exitCode=0 Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.857271 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"df8e056c06ac241c46a927c4e3f5a28f39d042bf13a8c672499d528786dbbbb0"} Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.858376 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"8f0b8a806330c630468951158113ee0ecaad5dbddb81b09c063752a17f21282c"} Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.858416 4948 scope.go:117] "RemoveContainer" containerID="b5405a03f834a339ac758e6efb1e31a8c094b3f6b973ea8540eefc716ba7bf7f" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.858594 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c08d-account-create-update-54vb7" event={"ID":"333ba21a-82a9-4f39-9ec5-afb7ae2ccf60","Type":"ContainerDied","Data":"7eca72d6b91c88e5a2e15a5744106cd7576ef511dae119df0c47c0377f312a2d"} Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.858637 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c08d-account-create-update-54vb7" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.858639 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7eca72d6b91c88e5a2e15a5744106cd7576ef511dae119df0c47c0377f312a2d" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.872919 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rnvbh" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.873085 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rnvbh" event={"ID":"92714d74-2a44-42e9-8829-d99ac773ac75","Type":"ContainerDied","Data":"5cee8fe5e9ab322142fa63fb214c8bbff5f35bb77dce1e1e2cae0a5fb25cc1c8"} Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.873134 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cee8fe5e9ab322142fa63fb214c8bbff5f35bb77dce1e1e2cae0a5fb25cc1c8" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.874956 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"790894c9-053c-497c-955d-ce7519111dd6","Type":"ContainerStarted","Data":"ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a"} Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.875502 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.877185 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fs9rr" event={"ID":"4d815209-521c-4c7f-a026-18899832459f","Type":"ContainerDied","Data":"83bd7271243acafa6581de499ba5b4200de0d064e497f5bdb3745d092e153305"} Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.877217 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83bd7271243acafa6581de499ba5b4200de0d064e497f5bdb3745d092e153305" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.877218 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fs9rr" Feb 20 08:21:38 crc kubenswrapper[4948]: I0220 08:21:38.915250 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=55.915236221 podStartE2EDuration="55.915236221s" podCreationTimestamp="2026-02-20 08:20:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:38.914516603 +0000 UTC m=+947.889011453" watchObservedRunningTime="2026-02-20 08:21:38.915236221 +0000 UTC m=+947.889731041" Feb 20 08:21:39 crc kubenswrapper[4948]: I0220 08:21:39.849020 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:21:39 crc kubenswrapper[4948]: I0220 08:21:39.912950 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hmhzl"] Feb 20 08:21:39 crc kubenswrapper[4948]: I0220 08:21:39.913251 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" podUID="33a06f41-ea8b-4138-8048-5425cd04f015" containerName="dnsmasq-dns" containerID="cri-o://489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f" gracePeriod=10 Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.378172 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.422477 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-config\") pod \"33a06f41-ea8b-4138-8048-5425cd04f015\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.422529 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-dns-svc\") pod \"33a06f41-ea8b-4138-8048-5425cd04f015\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.422595 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-nb\") pod \"33a06f41-ea8b-4138-8048-5425cd04f015\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.422744 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2twp\" (UniqueName: \"kubernetes.io/projected/33a06f41-ea8b-4138-8048-5425cd04f015-kube-api-access-c2twp\") pod \"33a06f41-ea8b-4138-8048-5425cd04f015\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.422910 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-sb\") pod \"33a06f41-ea8b-4138-8048-5425cd04f015\" (UID: \"33a06f41-ea8b-4138-8048-5425cd04f015\") " Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.439889 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33a06f41-ea8b-4138-8048-5425cd04f015-kube-api-access-c2twp" (OuterVolumeSpecName: "kube-api-access-c2twp") pod "33a06f41-ea8b-4138-8048-5425cd04f015" (UID: "33a06f41-ea8b-4138-8048-5425cd04f015"). InnerVolumeSpecName "kube-api-access-c2twp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.458863 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-config" (OuterVolumeSpecName: "config") pod "33a06f41-ea8b-4138-8048-5425cd04f015" (UID: "33a06f41-ea8b-4138-8048-5425cd04f015"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.464527 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "33a06f41-ea8b-4138-8048-5425cd04f015" (UID: "33a06f41-ea8b-4138-8048-5425cd04f015"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.465764 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "33a06f41-ea8b-4138-8048-5425cd04f015" (UID: "33a06f41-ea8b-4138-8048-5425cd04f015"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.467496 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "33a06f41-ea8b-4138-8048-5425cd04f015" (UID: "33a06f41-ea8b-4138-8048-5425cd04f015"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.525394 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.525420 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.525429 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.525438 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33a06f41-ea8b-4138-8048-5425cd04f015-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.525446 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2twp\" (UniqueName: \"kubernetes.io/projected/33a06f41-ea8b-4138-8048-5425cd04f015-kube-api-access-c2twp\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.539324 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-rnvbh"] Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.544293 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-rnvbh"] Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.896203 4948 generic.go:334] "Generic (PLEG): container finished" podID="33a06f41-ea8b-4138-8048-5425cd04f015" containerID="489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f" exitCode=0 Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.896249 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" event={"ID":"33a06f41-ea8b-4138-8048-5425cd04f015","Type":"ContainerDied","Data":"489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f"} Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.896277 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" event={"ID":"33a06f41-ea8b-4138-8048-5425cd04f015","Type":"ContainerDied","Data":"1f7e5587a47afc9ffa01b7b40dc0afdae69e9922140a4a9e4162b6a2a7417196"} Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.896296 4948 scope.go:117] "RemoveContainer" containerID="489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.896395 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-hmhzl" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.953783 4948 scope.go:117] "RemoveContainer" containerID="664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.956098 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hmhzl"] Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.966526 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-hmhzl"] Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.979825 4948 scope.go:117] "RemoveContainer" containerID="489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f" Feb 20 08:21:40 crc kubenswrapper[4948]: E0220 08:21:40.980302 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f\": container with ID starting with 489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f not found: ID does not exist" containerID="489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.980345 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f"} err="failed to get container status \"489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f\": rpc error: code = NotFound desc = could not find container \"489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f\": container with ID starting with 489a052e4c0fb8182ef93444ec3a663e93256a532f88c70856200cb8dc8b254f not found: ID does not exist" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.980370 4948 scope.go:117] "RemoveContainer" containerID="664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a" Feb 20 08:21:40 crc kubenswrapper[4948]: E0220 08:21:40.980705 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a\": container with ID starting with 664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a not found: ID does not exist" containerID="664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a" Feb 20 08:21:40 crc kubenswrapper[4948]: I0220 08:21:40.980731 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a"} err="failed to get container status \"664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a\": rpc error: code = NotFound desc = could not find container \"664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a\": container with ID starting with 664c1ad747d8bd0a828811f06f4556fb8972cb9b47fe4b7d90fa3826c1e33b4a not found: ID does not exist" Feb 20 08:21:41 crc kubenswrapper[4948]: I0220 08:21:41.464209 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 20 08:21:41 crc kubenswrapper[4948]: I0220 08:21:41.731513 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33a06f41-ea8b-4138-8048-5425cd04f015" path="/var/lib/kubelet/pods/33a06f41-ea8b-4138-8048-5425cd04f015/volumes" Feb 20 08:21:41 crc kubenswrapper[4948]: I0220 08:21:41.732126 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92714d74-2a44-42e9-8829-d99ac773ac75" path="/var/lib/kubelet/pods/92714d74-2a44-42e9-8829-d99ac773ac75/volumes" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.511717 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-pkgzx"] Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512159 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a06f41-ea8b-4138-8048-5425cd04f015" containerName="dnsmasq-dns" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512203 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a06f41-ea8b-4138-8048-5425cd04f015" containerName="dnsmasq-dns" Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512223 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52238644-5964-427e-8b77-9ca228d031db" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512231 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="52238644-5964-427e-8b77-9ca228d031db" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512248 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512257 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512272 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="333ba21a-82a9-4f39-9ec5-afb7ae2ccf60" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512279 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="333ba21a-82a9-4f39-9ec5-afb7ae2ccf60" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512289 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f917fa7f-51f9-4c8f-bee2-11529a6d44a6" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512297 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f917fa7f-51f9-4c8f-bee2-11529a6d44a6" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512311 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d815209-521c-4c7f-a026-18899832459f" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512319 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d815209-521c-4c7f-a026-18899832459f" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512334 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92714d74-2a44-42e9-8829-d99ac773ac75" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512342 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="92714d74-2a44-42e9-8829-d99ac773ac75" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512356 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a06f41-ea8b-4138-8048-5425cd04f015" containerName="init" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512364 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a06f41-ea8b-4138-8048-5425cd04f015" containerName="init" Feb 20 08:21:42 crc kubenswrapper[4948]: E0220 08:21:42.512377 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5bf7a21-af5d-4334-bc98-bd0f9eb9035a" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512386 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5bf7a21-af5d-4334-bc98-bd0f9eb9035a" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512567 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f917fa7f-51f9-4c8f-bee2-11529a6d44a6" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512586 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512595 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5bf7a21-af5d-4334-bc98-bd0f9eb9035a" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512611 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="92714d74-2a44-42e9-8829-d99ac773ac75" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512620 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d815209-521c-4c7f-a026-18899832459f" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512636 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="33a06f41-ea8b-4138-8048-5425cd04f015" containerName="dnsmasq-dns" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512648 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="52238644-5964-427e-8b77-9ca228d031db" containerName="mariadb-database-create" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.512658 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="333ba21a-82a9-4f39-9ec5-afb7ae2ccf60" containerName="mariadb-account-create-update" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.513304 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.529576 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.534610 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bgxj8" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.543634 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-pkgzx"] Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.660512 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-db-sync-config-data\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.660571 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-combined-ca-bundle\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.660640 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-config-data\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.661378 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j589d\" (UniqueName: \"kubernetes.io/projected/65735728-e734-40de-a185-9caa1db9a47d-kube-api-access-j589d\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.762997 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-config-data\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.763080 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j589d\" (UniqueName: \"kubernetes.io/projected/65735728-e734-40de-a185-9caa1db9a47d-kube-api-access-j589d\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.763124 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-db-sync-config-data\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.763157 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-combined-ca-bundle\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.771908 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-combined-ca-bundle\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.778518 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-db-sync-config-data\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.778545 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-config-data\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.783678 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j589d\" (UniqueName: \"kubernetes.io/projected/65735728-e734-40de-a185-9caa1db9a47d-kube-api-access-j589d\") pod \"glance-db-sync-pkgzx\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.838138 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-pkgzx" Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.915019 4948 generic.go:334] "Generic (PLEG): container finished" podID="642adfd1-e253-4e0a-80e6-860a9f5c1ae1" containerID="8fbeb28cef63fcfd89707879381c2e4ed399e4426a2e6226acf2ce1d9b7e8e2c" exitCode=0 Feb 20 08:21:42 crc kubenswrapper[4948]: I0220 08:21:42.915072 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-dllls" event={"ID":"642adfd1-e253-4e0a-80e6-860a9f5c1ae1","Type":"ContainerDied","Data":"8fbeb28cef63fcfd89707879381c2e4ed399e4426a2e6226acf2ce1d9b7e8e2c"} Feb 20 08:21:43 crc kubenswrapper[4948]: I0220 08:21:43.389633 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-pkgzx"] Feb 20 08:21:43 crc kubenswrapper[4948]: W0220 08:21:43.398704 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65735728_e734_40de_a185_9caa1db9a47d.slice/crio-5e99cbf3924927a30ad54b19ba1a0b9e160811b4e81053363ade5b0662fa69fa WatchSource:0}: Error finding container 5e99cbf3924927a30ad54b19ba1a0b9e160811b4e81053363ade5b0662fa69fa: Status 404 returned error can't find the container with id 5e99cbf3924927a30ad54b19ba1a0b9e160811b4e81053363ade5b0662fa69fa Feb 20 08:21:43 crc kubenswrapper[4948]: I0220 08:21:43.925186 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-pkgzx" event={"ID":"65735728-e734-40de-a185-9caa1db9a47d","Type":"ContainerStarted","Data":"5e99cbf3924927a30ad54b19ba1a0b9e160811b4e81053363ade5b0662fa69fa"} Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.231696 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-n7gxj"] Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.252319 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.261105 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.262630 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-n7gxj"] Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.287055 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.287275 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538046ba-61c4-43ff-ad1f-a8d80d2ab075-operator-scripts\") pod \"root-account-create-update-n7gxj\" (UID: \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\") " pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.287330 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxckc\" (UniqueName: \"kubernetes.io/projected/538046ba-61c4-43ff-ad1f-a8d80d2ab075-kube-api-access-jxckc\") pod \"root-account-create-update-n7gxj\" (UID: \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\") " pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389091 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-scripts\") pod \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389132 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-combined-ca-bundle\") pod \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389197 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-dispersionconf\") pod \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389223 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-swiftconf\") pod \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389299 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-ring-data-devices\") pod \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389335 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-etc-swift\") pod \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389362 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89jgn\" (UniqueName: \"kubernetes.io/projected/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-kube-api-access-89jgn\") pod \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\" (UID: \"642adfd1-e253-4e0a-80e6-860a9f5c1ae1\") " Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389580 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538046ba-61c4-43ff-ad1f-a8d80d2ab075-operator-scripts\") pod \"root-account-create-update-n7gxj\" (UID: \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\") " pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.389608 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxckc\" (UniqueName: \"kubernetes.io/projected/538046ba-61c4-43ff-ad1f-a8d80d2ab075-kube-api-access-jxckc\") pod \"root-account-create-update-n7gxj\" (UID: \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\") " pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.390339 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "642adfd1-e253-4e0a-80e6-860a9f5c1ae1" (UID: "642adfd1-e253-4e0a-80e6-860a9f5c1ae1"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.390395 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "642adfd1-e253-4e0a-80e6-860a9f5c1ae1" (UID: "642adfd1-e253-4e0a-80e6-860a9f5c1ae1"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.390633 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538046ba-61c4-43ff-ad1f-a8d80d2ab075-operator-scripts\") pod \"root-account-create-update-n7gxj\" (UID: \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\") " pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.395504 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-kube-api-access-89jgn" (OuterVolumeSpecName: "kube-api-access-89jgn") pod "642adfd1-e253-4e0a-80e6-860a9f5c1ae1" (UID: "642adfd1-e253-4e0a-80e6-860a9f5c1ae1"). InnerVolumeSpecName "kube-api-access-89jgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.398277 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "642adfd1-e253-4e0a-80e6-860a9f5c1ae1" (UID: "642adfd1-e253-4e0a-80e6-860a9f5c1ae1"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.406491 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxckc\" (UniqueName: \"kubernetes.io/projected/538046ba-61c4-43ff-ad1f-a8d80d2ab075-kube-api-access-jxckc\") pod \"root-account-create-update-n7gxj\" (UID: \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\") " pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.416814 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "642adfd1-e253-4e0a-80e6-860a9f5c1ae1" (UID: "642adfd1-e253-4e0a-80e6-860a9f5c1ae1"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.417088 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-scripts" (OuterVolumeSpecName: "scripts") pod "642adfd1-e253-4e0a-80e6-860a9f5c1ae1" (UID: "642adfd1-e253-4e0a-80e6-860a9f5c1ae1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.424927 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "642adfd1-e253-4e0a-80e6-860a9f5c1ae1" (UID: "642adfd1-e253-4e0a-80e6-860a9f5c1ae1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.490350 4948 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.490383 4948 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.490393 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89jgn\" (UniqueName: \"kubernetes.io/projected/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-kube-api-access-89jgn\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.490404 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.490413 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.490421 4948 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.490431 4948 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/642adfd1-e253-4e0a-80e6-860a9f5c1ae1-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.598195 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.934631 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-dllls" event={"ID":"642adfd1-e253-4e0a-80e6-860a9f5c1ae1","Type":"ContainerDied","Data":"befd58899320f920bf3b69e7a9c823d94bc5bd3a3193731f420fe312a97c42e6"} Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.934669 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="befd58899320f920bf3b69e7a9c823d94bc5bd3a3193731f420fe312a97c42e6" Feb 20 08:21:44 crc kubenswrapper[4948]: I0220 08:21:44.934700 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dllls" Feb 20 08:21:45 crc kubenswrapper[4948]: I0220 08:21:45.059680 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-n7gxj"] Feb 20 08:21:45 crc kubenswrapper[4948]: W0220 08:21:45.064233 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod538046ba_61c4_43ff_ad1f_a8d80d2ab075.slice/crio-61b1645ad6199f5a1d388077d4b580c92f8d45bff4fda65d7f5262e0a615940a WatchSource:0}: Error finding container 61b1645ad6199f5a1d388077d4b580c92f8d45bff4fda65d7f5262e0a615940a: Status 404 returned error can't find the container with id 61b1645ad6199f5a1d388077d4b580c92f8d45bff4fda65d7f5262e0a615940a Feb 20 08:21:45 crc kubenswrapper[4948]: I0220 08:21:45.944143 4948 generic.go:334] "Generic (PLEG): container finished" podID="538046ba-61c4-43ff-ad1f-a8d80d2ab075" containerID="c4af4c0a8310dde22c0962922d9e1a2de48ad4ce8cecfbf27d7c0a78b93a2059" exitCode=0 Feb 20 08:21:45 crc kubenswrapper[4948]: I0220 08:21:45.944280 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-n7gxj" event={"ID":"538046ba-61c4-43ff-ad1f-a8d80d2ab075","Type":"ContainerDied","Data":"c4af4c0a8310dde22c0962922d9e1a2de48ad4ce8cecfbf27d7c0a78b93a2059"} Feb 20 08:21:45 crc kubenswrapper[4948]: I0220 08:21:45.944430 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-n7gxj" event={"ID":"538046ba-61c4-43ff-ad1f-a8d80d2ab075","Type":"ContainerStarted","Data":"61b1645ad6199f5a1d388077d4b580c92f8d45bff4fda65d7f5262e0a615940a"} Feb 20 08:21:46 crc kubenswrapper[4948]: I0220 08:21:46.543774 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:46 crc kubenswrapper[4948]: I0220 08:21:46.553421 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4d60def-5dc9-460b-92ed-e3d0157982ec-etc-swift\") pod \"swift-storage-0\" (UID: \"e4d60def-5dc9-460b-92ed-e3d0157982ec\") " pod="openstack/swift-storage-0" Feb 20 08:21:46 crc kubenswrapper[4948]: I0220 08:21:46.581332 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.092253 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.208243 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-mkpng" podUID="b80cb988-de2a-4e65-9161-a0af0561c754" containerName="ovn-controller" probeResult="failure" output=< Feb 20 08:21:47 crc kubenswrapper[4948]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 20 08:21:47 crc kubenswrapper[4948]: > Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.238603 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.244142 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-q4bv4" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.288632 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.476407 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxckc\" (UniqueName: \"kubernetes.io/projected/538046ba-61c4-43ff-ad1f-a8d80d2ab075-kube-api-access-jxckc\") pod \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\" (UID: \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\") " Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.476554 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538046ba-61c4-43ff-ad1f-a8d80d2ab075-operator-scripts\") pod \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\" (UID: \"538046ba-61c4-43ff-ad1f-a8d80d2ab075\") " Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.477366 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/538046ba-61c4-43ff-ad1f-a8d80d2ab075-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "538046ba-61c4-43ff-ad1f-a8d80d2ab075" (UID: "538046ba-61c4-43ff-ad1f-a8d80d2ab075"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.484300 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/538046ba-61c4-43ff-ad1f-a8d80d2ab075-kube-api-access-jxckc" (OuterVolumeSpecName: "kube-api-access-jxckc") pod "538046ba-61c4-43ff-ad1f-a8d80d2ab075" (UID: "538046ba-61c4-43ff-ad1f-a8d80d2ab075"). InnerVolumeSpecName "kube-api-access-jxckc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.491334 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-mkpng-config-g4w5d"] Feb 20 08:21:47 crc kubenswrapper[4948]: E0220 08:21:47.491692 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642adfd1-e253-4e0a-80e6-860a9f5c1ae1" containerName="swift-ring-rebalance" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.491705 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="642adfd1-e253-4e0a-80e6-860a9f5c1ae1" containerName="swift-ring-rebalance" Feb 20 08:21:47 crc kubenswrapper[4948]: E0220 08:21:47.491717 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="538046ba-61c4-43ff-ad1f-a8d80d2ab075" containerName="mariadb-account-create-update" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.491722 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="538046ba-61c4-43ff-ad1f-a8d80d2ab075" containerName="mariadb-account-create-update" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.493396 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="642adfd1-e253-4e0a-80e6-860a9f5c1ae1" containerName="swift-ring-rebalance" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.493417 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="538046ba-61c4-43ff-ad1f-a8d80d2ab075" containerName="mariadb-account-create-update" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.494022 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.497952 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.557995 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mkpng-config-g4w5d"] Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.578973 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.579076 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run-ovn\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.579096 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-scripts\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.579117 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-log-ovn\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.579197 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8667m\" (UniqueName: \"kubernetes.io/projected/fa672fa0-b98b-477c-a34c-ce996ce6592d-kube-api-access-8667m\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.579322 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-additional-scripts\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.579517 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxckc\" (UniqueName: \"kubernetes.io/projected/538046ba-61c4-43ff-ad1f-a8d80d2ab075-kube-api-access-jxckc\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.579542 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538046ba-61c4-43ff-ad1f-a8d80d2ab075-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.681544 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.681706 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run-ovn\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.681730 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-scripts\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.681762 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-log-ovn\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.681758 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.681832 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run-ovn\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.681922 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8667m\" (UniqueName: \"kubernetes.io/projected/fa672fa0-b98b-477c-a34c-ce996ce6592d-kube-api-access-8667m\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.681971 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-additional-scripts\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.683039 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-additional-scripts\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.683098 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-log-ovn\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.684140 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-scripts\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.700230 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8667m\" (UniqueName: \"kubernetes.io/projected/fa672fa0-b98b-477c-a34c-ce996ce6592d-kube-api-access-8667m\") pod \"ovn-controller-mkpng-config-g4w5d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.900163 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.963791 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"715d365ca5f7dd84a91d85b1e876e17deb71037af959f60dad725383a0ccf801"} Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.965965 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-n7gxj" Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.965980 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-n7gxj" event={"ID":"538046ba-61c4-43ff-ad1f-a8d80d2ab075","Type":"ContainerDied","Data":"61b1645ad6199f5a1d388077d4b580c92f8d45bff4fda65d7f5262e0a615940a"} Feb 20 08:21:47 crc kubenswrapper[4948]: I0220 08:21:47.966147 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61b1645ad6199f5a1d388077d4b580c92f8d45bff4fda65d7f5262e0a615940a" Feb 20 08:21:48 crc kubenswrapper[4948]: I0220 08:21:48.470441 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mkpng-config-g4w5d"] Feb 20 08:21:48 crc kubenswrapper[4948]: W0220 08:21:48.479836 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa672fa0_b98b_477c_a34c_ce996ce6592d.slice/crio-06f713018be060e56a2397e33a80f4b214dc9b77719605af6934b96eaf9c8488 WatchSource:0}: Error finding container 06f713018be060e56a2397e33a80f4b214dc9b77719605af6934b96eaf9c8488: Status 404 returned error can't find the container with id 06f713018be060e56a2397e33a80f4b214dc9b77719605af6934b96eaf9c8488 Feb 20 08:21:48 crc kubenswrapper[4948]: I0220 08:21:48.973400 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mkpng-config-g4w5d" event={"ID":"fa672fa0-b98b-477c-a34c-ce996ce6592d","Type":"ContainerStarted","Data":"d4c99094574230bbe4d4f5bd6cf1019d664fb9f16355c0c0e8d85cd644e73d32"} Feb 20 08:21:48 crc kubenswrapper[4948]: I0220 08:21:48.973712 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mkpng-config-g4w5d" event={"ID":"fa672fa0-b98b-477c-a34c-ce996ce6592d","Type":"ContainerStarted","Data":"06f713018be060e56a2397e33a80f4b214dc9b77719605af6934b96eaf9c8488"} Feb 20 08:21:48 crc kubenswrapper[4948]: I0220 08:21:48.976364 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"1c61e52c45386cf02be5fb7e55dbb4aced10eac17718db550b046809d65a5500"} Feb 20 08:21:48 crc kubenswrapper[4948]: I0220 08:21:48.976387 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"10af57300e90334431370f0d766ad2b7c89c11757c0eec9308a6b68829956b58"} Feb 20 08:21:48 crc kubenswrapper[4948]: I0220 08:21:48.976396 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"0ff01f06ef84d4169843f5686becda94c1c31051a153f7883fa0db15d0864b19"} Feb 20 08:21:48 crc kubenswrapper[4948]: I0220 08:21:48.976405 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"ab55b41ba892527d39efc04b2e4d4f30976ca85920ec9c5a5339f15470e49c93"} Feb 20 08:21:48 crc kubenswrapper[4948]: I0220 08:21:48.990381 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-mkpng-config-g4w5d" podStartSLOduration=1.9903651199999999 podStartE2EDuration="1.99036512s" podCreationTimestamp="2026-02-20 08:21:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:48.989530439 +0000 UTC m=+957.964025259" watchObservedRunningTime="2026-02-20 08:21:48.99036512 +0000 UTC m=+957.964859940" Feb 20 08:21:49 crc kubenswrapper[4948]: I0220 08:21:49.989985 4948 generic.go:334] "Generic (PLEG): container finished" podID="fa672fa0-b98b-477c-a34c-ce996ce6592d" containerID="d4c99094574230bbe4d4f5bd6cf1019d664fb9f16355c0c0e8d85cd644e73d32" exitCode=0 Feb 20 08:21:49 crc kubenswrapper[4948]: I0220 08:21:49.990037 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mkpng-config-g4w5d" event={"ID":"fa672fa0-b98b-477c-a34c-ce996ce6592d","Type":"ContainerDied","Data":"d4c99094574230bbe4d4f5bd6cf1019d664fb9f16355c0c0e8d85cd644e73d32"} Feb 20 08:21:50 crc kubenswrapper[4948]: I0220 08:21:50.546638 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-n7gxj"] Feb 20 08:21:50 crc kubenswrapper[4948]: I0220 08:21:50.556057 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-n7gxj"] Feb 20 08:21:51 crc kubenswrapper[4948]: I0220 08:21:51.745165 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="538046ba-61c4-43ff-ad1f-a8d80d2ab075" path="/var/lib/kubelet/pods/538046ba-61c4-43ff-ad1f-a8d80d2ab075/volumes" Feb 20 08:21:52 crc kubenswrapper[4948]: I0220 08:21:52.209811 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-mkpng" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.121276 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.460078 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-97j89"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.461604 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-97j89" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.471355 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-97j89"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.503164 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.568149 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-0b2f-account-create-update-2f74q"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.569396 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.571548 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.585371 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-0b2f-account-create-update-2f74q"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.609732 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ced4ac5-e057-476a-908a-f573a8cf997b-operator-scripts\") pod \"cinder-db-create-97j89\" (UID: \"4ced4ac5-e057-476a-908a-f573a8cf997b\") " pod="openstack/cinder-db-create-97j89" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.609917 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrc6c\" (UniqueName: \"kubernetes.io/projected/4ced4ac5-e057-476a-908a-f573a8cf997b-kube-api-access-rrc6c\") pod \"cinder-db-create-97j89\" (UID: \"4ced4ac5-e057-476a-908a-f573a8cf997b\") " pod="openstack/cinder-db-create-97j89" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.644283 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-4djkr"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.649345 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.659698 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4djkr"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.698624 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-a86b-account-create-update-mspp7"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.699541 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.701185 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.710799 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6d068f3-210b-41e8-b90a-5406759c0606-operator-scripts\") pod \"cinder-0b2f-account-create-update-2f74q\" (UID: \"b6d068f3-210b-41e8-b90a-5406759c0606\") " pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.711003 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrc6c\" (UniqueName: \"kubernetes.io/projected/4ced4ac5-e057-476a-908a-f573a8cf997b-kube-api-access-rrc6c\") pod \"cinder-db-create-97j89\" (UID: \"4ced4ac5-e057-476a-908a-f573a8cf997b\") " pod="openstack/cinder-db-create-97j89" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.711101 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ced4ac5-e057-476a-908a-f573a8cf997b-operator-scripts\") pod \"cinder-db-create-97j89\" (UID: \"4ced4ac5-e057-476a-908a-f573a8cf997b\") " pod="openstack/cinder-db-create-97j89" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.711217 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c86dh\" (UniqueName: \"kubernetes.io/projected/b6d068f3-210b-41e8-b90a-5406759c0606-kube-api-access-c86dh\") pod \"cinder-0b2f-account-create-update-2f74q\" (UID: \"b6d068f3-210b-41e8-b90a-5406759c0606\") " pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.712122 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ced4ac5-e057-476a-908a-f573a8cf997b-operator-scripts\") pod \"cinder-db-create-97j89\" (UID: \"4ced4ac5-e057-476a-908a-f573a8cf997b\") " pod="openstack/cinder-db-create-97j89" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.723829 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a86b-account-create-update-mspp7"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.738032 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrc6c\" (UniqueName: \"kubernetes.io/projected/4ced4ac5-e057-476a-908a-f573a8cf997b-kube-api-access-rrc6c\") pod \"cinder-db-create-97j89\" (UID: \"4ced4ac5-e057-476a-908a-f573a8cf997b\") " pod="openstack/cinder-db-create-97j89" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.758219 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-w4bzp"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.763478 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-w4bzp" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.772026 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-w4bzp"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.785972 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-97j89" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.812759 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10eedd01-3f52-4d44-981d-27bc16af186b-operator-scripts\") pod \"barbican-db-create-4djkr\" (UID: \"10eedd01-3f52-4d44-981d-27bc16af186b\") " pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.812813 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3f86085-113f-408e-b43c-a499e2fe8b90-operator-scripts\") pod \"barbican-a86b-account-create-update-mspp7\" (UID: \"d3f86085-113f-408e-b43c-a499e2fe8b90\") " pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.812877 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k984h\" (UniqueName: \"kubernetes.io/projected/d3f86085-113f-408e-b43c-a499e2fe8b90-kube-api-access-k984h\") pod \"barbican-a86b-account-create-update-mspp7\" (UID: \"d3f86085-113f-408e-b43c-a499e2fe8b90\") " pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.812919 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c86dh\" (UniqueName: \"kubernetes.io/projected/b6d068f3-210b-41e8-b90a-5406759c0606-kube-api-access-c86dh\") pod \"cinder-0b2f-account-create-update-2f74q\" (UID: \"b6d068f3-210b-41e8-b90a-5406759c0606\") " pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.812959 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6d068f3-210b-41e8-b90a-5406759c0606-operator-scripts\") pod \"cinder-0b2f-account-create-update-2f74q\" (UID: \"b6d068f3-210b-41e8-b90a-5406759c0606\") " pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.813038 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr2kf\" (UniqueName: \"kubernetes.io/projected/10eedd01-3f52-4d44-981d-27bc16af186b-kube-api-access-kr2kf\") pod \"barbican-db-create-4djkr\" (UID: \"10eedd01-3f52-4d44-981d-27bc16af186b\") " pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.813956 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6d068f3-210b-41e8-b90a-5406759c0606-operator-scripts\") pod \"cinder-0b2f-account-create-update-2f74q\" (UID: \"b6d068f3-210b-41e8-b90a-5406759c0606\") " pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.837164 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-gz5j7"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.838201 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c86dh\" (UniqueName: \"kubernetes.io/projected/b6d068f3-210b-41e8-b90a-5406759c0606-kube-api-access-c86dh\") pod \"cinder-0b2f-account-create-update-2f74q\" (UID: \"b6d068f3-210b-41e8-b90a-5406759c0606\") " pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.838766 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.849295 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.849580 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.849685 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.849827 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-v9rj2" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.858941 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gz5j7"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.884546 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-e6d0-account-create-update-cdxd5"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.886134 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.890035 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.895839 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-e6d0-account-create-update-cdxd5"] Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.904806 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.913969 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr2kf\" (UniqueName: \"kubernetes.io/projected/10eedd01-3f52-4d44-981d-27bc16af186b-kube-api-access-kr2kf\") pod \"barbican-db-create-4djkr\" (UID: \"10eedd01-3f52-4d44-981d-27bc16af186b\") " pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.914023 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/501e504c-7910-47cb-a181-812a0d72f2db-operator-scripts\") pod \"neutron-db-create-w4bzp\" (UID: \"501e504c-7910-47cb-a181-812a0d72f2db\") " pod="openstack/neutron-db-create-w4bzp" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.914069 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10eedd01-3f52-4d44-981d-27bc16af186b-operator-scripts\") pod \"barbican-db-create-4djkr\" (UID: \"10eedd01-3f52-4d44-981d-27bc16af186b\") " pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.914091 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3f86085-113f-408e-b43c-a499e2fe8b90-operator-scripts\") pod \"barbican-a86b-account-create-update-mspp7\" (UID: \"d3f86085-113f-408e-b43c-a499e2fe8b90\") " pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.914242 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k984h\" (UniqueName: \"kubernetes.io/projected/d3f86085-113f-408e-b43c-a499e2fe8b90-kube-api-access-k984h\") pod \"barbican-a86b-account-create-update-mspp7\" (UID: \"d3f86085-113f-408e-b43c-a499e2fe8b90\") " pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.914311 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvqt6\" (UniqueName: \"kubernetes.io/projected/501e504c-7910-47cb-a181-812a0d72f2db-kube-api-access-bvqt6\") pod \"neutron-db-create-w4bzp\" (UID: \"501e504c-7910-47cb-a181-812a0d72f2db\") " pod="openstack/neutron-db-create-w4bzp" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.914711 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3f86085-113f-408e-b43c-a499e2fe8b90-operator-scripts\") pod \"barbican-a86b-account-create-update-mspp7\" (UID: \"d3f86085-113f-408e-b43c-a499e2fe8b90\") " pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.914834 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10eedd01-3f52-4d44-981d-27bc16af186b-operator-scripts\") pod \"barbican-db-create-4djkr\" (UID: \"10eedd01-3f52-4d44-981d-27bc16af186b\") " pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.932008 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr2kf\" (UniqueName: \"kubernetes.io/projected/10eedd01-3f52-4d44-981d-27bc16af186b-kube-api-access-kr2kf\") pod \"barbican-db-create-4djkr\" (UID: \"10eedd01-3f52-4d44-981d-27bc16af186b\") " pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.933647 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k984h\" (UniqueName: \"kubernetes.io/projected/d3f86085-113f-408e-b43c-a499e2fe8b90-kube-api-access-k984h\") pod \"barbican-a86b-account-create-update-mspp7\" (UID: \"d3f86085-113f-408e-b43c-a499e2fe8b90\") " pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:54 crc kubenswrapper[4948]: I0220 08:21:54.967425 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.016586 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-combined-ca-bundle\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.016696 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-config-data\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.016730 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx8w6\" (UniqueName: \"kubernetes.io/projected/b9370de8-8b72-4819-af79-4a9ee1fb6777-kube-api-access-zx8w6\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.016761 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvqt6\" (UniqueName: \"kubernetes.io/projected/501e504c-7910-47cb-a181-812a0d72f2db-kube-api-access-bvqt6\") pod \"neutron-db-create-w4bzp\" (UID: \"501e504c-7910-47cb-a181-812a0d72f2db\") " pod="openstack/neutron-db-create-w4bzp" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.016799 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a739e4d-8960-46f3-8c5d-2373b59165dd-operator-scripts\") pod \"neutron-e6d0-account-create-update-cdxd5\" (UID: \"0a739e4d-8960-46f3-8c5d-2373b59165dd\") " pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.016857 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dqq5\" (UniqueName: \"kubernetes.io/projected/0a739e4d-8960-46f3-8c5d-2373b59165dd-kube-api-access-2dqq5\") pod \"neutron-e6d0-account-create-update-cdxd5\" (UID: \"0a739e4d-8960-46f3-8c5d-2373b59165dd\") " pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.017020 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/501e504c-7910-47cb-a181-812a0d72f2db-operator-scripts\") pod \"neutron-db-create-w4bzp\" (UID: \"501e504c-7910-47cb-a181-812a0d72f2db\") " pod="openstack/neutron-db-create-w4bzp" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.017815 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/501e504c-7910-47cb-a181-812a0d72f2db-operator-scripts\") pod \"neutron-db-create-w4bzp\" (UID: \"501e504c-7910-47cb-a181-812a0d72f2db\") " pod="openstack/neutron-db-create-w4bzp" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.030385 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.038597 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvqt6\" (UniqueName: \"kubernetes.io/projected/501e504c-7910-47cb-a181-812a0d72f2db-kube-api-access-bvqt6\") pod \"neutron-db-create-w4bzp\" (UID: \"501e504c-7910-47cb-a181-812a0d72f2db\") " pod="openstack/neutron-db-create-w4bzp" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.095438 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-w4bzp" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.118577 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-combined-ca-bundle\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.118640 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-config-data\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.118659 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx8w6\" (UniqueName: \"kubernetes.io/projected/b9370de8-8b72-4819-af79-4a9ee1fb6777-kube-api-access-zx8w6\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.118684 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a739e4d-8960-46f3-8c5d-2373b59165dd-operator-scripts\") pod \"neutron-e6d0-account-create-update-cdxd5\" (UID: \"0a739e4d-8960-46f3-8c5d-2373b59165dd\") " pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.118724 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dqq5\" (UniqueName: \"kubernetes.io/projected/0a739e4d-8960-46f3-8c5d-2373b59165dd-kube-api-access-2dqq5\") pod \"neutron-e6d0-account-create-update-cdxd5\" (UID: \"0a739e4d-8960-46f3-8c5d-2373b59165dd\") " pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.119827 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a739e4d-8960-46f3-8c5d-2373b59165dd-operator-scripts\") pod \"neutron-e6d0-account-create-update-cdxd5\" (UID: \"0a739e4d-8960-46f3-8c5d-2373b59165dd\") " pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.122625 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-config-data\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.122781 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-combined-ca-bundle\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.137770 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dqq5\" (UniqueName: \"kubernetes.io/projected/0a739e4d-8960-46f3-8c5d-2373b59165dd-kube-api-access-2dqq5\") pod \"neutron-e6d0-account-create-update-cdxd5\" (UID: \"0a739e4d-8960-46f3-8c5d-2373b59165dd\") " pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.140724 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx8w6\" (UniqueName: \"kubernetes.io/projected/b9370de8-8b72-4819-af79-4a9ee1fb6777-kube-api-access-zx8w6\") pod \"keystone-db-sync-gz5j7\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.184902 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.199862 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.558389 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-zvzvk"] Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.559330 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-zvzvk" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.564400 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.589579 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-zvzvk"] Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.727940 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcv9f\" (UniqueName: \"kubernetes.io/projected/12f1f665-7a04-4794-a727-cc5c060a1933-kube-api-access-rcv9f\") pod \"root-account-create-update-zvzvk\" (UID: \"12f1f665-7a04-4794-a727-cc5c060a1933\") " pod="openstack/root-account-create-update-zvzvk" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.728035 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12f1f665-7a04-4794-a727-cc5c060a1933-operator-scripts\") pod \"root-account-create-update-zvzvk\" (UID: \"12f1f665-7a04-4794-a727-cc5c060a1933\") " pod="openstack/root-account-create-update-zvzvk" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.829870 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcv9f\" (UniqueName: \"kubernetes.io/projected/12f1f665-7a04-4794-a727-cc5c060a1933-kube-api-access-rcv9f\") pod \"root-account-create-update-zvzvk\" (UID: \"12f1f665-7a04-4794-a727-cc5c060a1933\") " pod="openstack/root-account-create-update-zvzvk" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.830163 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12f1f665-7a04-4794-a727-cc5c060a1933-operator-scripts\") pod \"root-account-create-update-zvzvk\" (UID: \"12f1f665-7a04-4794-a727-cc5c060a1933\") " pod="openstack/root-account-create-update-zvzvk" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.830842 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12f1f665-7a04-4794-a727-cc5c060a1933-operator-scripts\") pod \"root-account-create-update-zvzvk\" (UID: \"12f1f665-7a04-4794-a727-cc5c060a1933\") " pod="openstack/root-account-create-update-zvzvk" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.845152 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcv9f\" (UniqueName: \"kubernetes.io/projected/12f1f665-7a04-4794-a727-cc5c060a1933-kube-api-access-rcv9f\") pod \"root-account-create-update-zvzvk\" (UID: \"12f1f665-7a04-4794-a727-cc5c060a1933\") " pod="openstack/root-account-create-update-zvzvk" Feb 20 08:21:55 crc kubenswrapper[4948]: I0220 08:21:55.897524 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-zvzvk" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.547263 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.640159 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8667m\" (UniqueName: \"kubernetes.io/projected/fa672fa0-b98b-477c-a34c-ce996ce6592d-kube-api-access-8667m\") pod \"fa672fa0-b98b-477c-a34c-ce996ce6592d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.641845 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-additional-scripts\") pod \"fa672fa0-b98b-477c-a34c-ce996ce6592d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.641914 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-scripts\") pod \"fa672fa0-b98b-477c-a34c-ce996ce6592d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.642044 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run-ovn\") pod \"fa672fa0-b98b-477c-a34c-ce996ce6592d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.642083 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-log-ovn\") pod \"fa672fa0-b98b-477c-a34c-ce996ce6592d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.642113 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run\") pod \"fa672fa0-b98b-477c-a34c-ce996ce6592d\" (UID: \"fa672fa0-b98b-477c-a34c-ce996ce6592d\") " Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.642543 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run" (OuterVolumeSpecName: "var-run") pod "fa672fa0-b98b-477c-a34c-ce996ce6592d" (UID: "fa672fa0-b98b-477c-a34c-ce996ce6592d"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.643151 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "fa672fa0-b98b-477c-a34c-ce996ce6592d" (UID: "fa672fa0-b98b-477c-a34c-ce996ce6592d"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.643569 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "fa672fa0-b98b-477c-a34c-ce996ce6592d" (UID: "fa672fa0-b98b-477c-a34c-ce996ce6592d"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.643625 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "fa672fa0-b98b-477c-a34c-ce996ce6592d" (UID: "fa672fa0-b98b-477c-a34c-ce996ce6592d"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.644081 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-scripts" (OuterVolumeSpecName: "scripts") pod "fa672fa0-b98b-477c-a34c-ce996ce6592d" (UID: "fa672fa0-b98b-477c-a34c-ce996ce6592d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.649641 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa672fa0-b98b-477c-a34c-ce996ce6592d-kube-api-access-8667m" (OuterVolumeSpecName: "kube-api-access-8667m") pod "fa672fa0-b98b-477c-a34c-ce996ce6592d" (UID: "fa672fa0-b98b-477c-a34c-ce996ce6592d"). InnerVolumeSpecName "kube-api-access-8667m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.746766 4948 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.746798 4948 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.746809 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8667m\" (UniqueName: \"kubernetes.io/projected/fa672fa0-b98b-477c-a34c-ce996ce6592d-kube-api-access-8667m\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.746821 4948 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.746829 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa672fa0-b98b-477c-a34c-ce996ce6592d-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:56 crc kubenswrapper[4948]: I0220 08:21:56.746836 4948 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa672fa0-b98b-477c-a34c-ce996ce6592d-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.064251 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mkpng-config-g4w5d" event={"ID":"fa672fa0-b98b-477c-a34c-ce996ce6592d","Type":"ContainerDied","Data":"06f713018be060e56a2397e33a80f4b214dc9b77719605af6934b96eaf9c8488"} Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.064488 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06f713018be060e56a2397e33a80f4b214dc9b77719605af6934b96eaf9c8488" Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.064547 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mkpng-config-g4w5d" Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.148661 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a86b-account-create-update-mspp7"] Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.382270 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-zvzvk"] Feb 20 08:21:57 crc kubenswrapper[4948]: W0220 08:21:57.385741 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12f1f665_7a04_4794_a727_cc5c060a1933.slice/crio-4ff5b9d3c7487921f2af53697fe4d63849f3213b42b4935658e6dfe3c2d2fd6c WatchSource:0}: Error finding container 4ff5b9d3c7487921f2af53697fe4d63849f3213b42b4935658e6dfe3c2d2fd6c: Status 404 returned error can't find the container with id 4ff5b9d3c7487921f2af53697fe4d63849f3213b42b4935658e6dfe3c2d2fd6c Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.387250 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-e6d0-account-create-update-cdxd5"] Feb 20 08:21:57 crc kubenswrapper[4948]: W0220 08:21:57.391412 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a739e4d_8960_46f3_8c5d_2373b59165dd.slice/crio-1e6ec887d9d3e16f02f5b4d5f7d7d830aca0b26b2e6976000ecd8e9ec7ae4ac5 WatchSource:0}: Error finding container 1e6ec887d9d3e16f02f5b4d5f7d7d830aca0b26b2e6976000ecd8e9ec7ae4ac5: Status 404 returned error can't find the container with id 1e6ec887d9d3e16f02f5b4d5f7d7d830aca0b26b2e6976000ecd8e9ec7ae4ac5 Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.400839 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4djkr"] Feb 20 08:21:57 crc kubenswrapper[4948]: W0220 08:21:57.424416 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10eedd01_3f52_4d44_981d_27bc16af186b.slice/crio-f4a16b89f1881652510d924497279d30d07a12e1a23c9a2a09460395d54b6754 WatchSource:0}: Error finding container f4a16b89f1881652510d924497279d30d07a12e1a23c9a2a09460395d54b6754: Status 404 returned error can't find the container with id f4a16b89f1881652510d924497279d30d07a12e1a23c9a2a09460395d54b6754 Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.583408 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-97j89"] Feb 20 08:21:57 crc kubenswrapper[4948]: W0220 08:21:57.593083 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ced4ac5_e057_476a_908a_f573a8cf997b.slice/crio-74d816dd0d8f2d728e06f6752ea62cf407348186231cd23345763d2a41ae00ad WatchSource:0}: Error finding container 74d816dd0d8f2d728e06f6752ea62cf407348186231cd23345763d2a41ae00ad: Status 404 returned error can't find the container with id 74d816dd0d8f2d728e06f6752ea62cf407348186231cd23345763d2a41ae00ad Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.597933 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-0b2f-account-create-update-2f74q"] Feb 20 08:21:57 crc kubenswrapper[4948]: W0220 08:21:57.606597 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6d068f3_210b_41e8_b90a_5406759c0606.slice/crio-67031a60eeb343261bf69ab0502142ef436d8c88a1348d9a5bdd6561d7700913 WatchSource:0}: Error finding container 67031a60eeb343261bf69ab0502142ef436d8c88a1348d9a5bdd6561d7700913: Status 404 returned error can't find the container with id 67031a60eeb343261bf69ab0502142ef436d8c88a1348d9a5bdd6561d7700913 Feb 20 08:21:57 crc kubenswrapper[4948]: W0220 08:21:57.612107 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9370de8_8b72_4819_af79_4a9ee1fb6777.slice/crio-fd31b197071ac286c0663672b75ab91e334d6c66ca4e2bf75a86eb8ab8c0137c WatchSource:0}: Error finding container fd31b197071ac286c0663672b75ab91e334d6c66ca4e2bf75a86eb8ab8c0137c: Status 404 returned error can't find the container with id fd31b197071ac286c0663672b75ab91e334d6c66ca4e2bf75a86eb8ab8c0137c Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.617069 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-w4bzp"] Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.644590 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gz5j7"] Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.682789 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-mkpng-config-g4w5d"] Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.690001 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-mkpng-config-g4w5d"] Feb 20 08:21:57 crc kubenswrapper[4948]: I0220 08:21:57.799805 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa672fa0-b98b-477c-a34c-ce996ce6592d" path="/var/lib/kubelet/pods/fa672fa0-b98b-477c-a34c-ce996ce6592d/volumes" Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.075516 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-zvzvk" event={"ID":"12f1f665-7a04-4794-a727-cc5c060a1933","Type":"ContainerStarted","Data":"cddaaa0c08e72fa702328c57550853b538c4537573467ac0ebd0761e1fcbc30c"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.075781 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-zvzvk" event={"ID":"12f1f665-7a04-4794-a727-cc5c060a1933","Type":"ContainerStarted","Data":"4ff5b9d3c7487921f2af53697fe4d63849f3213b42b4935658e6dfe3c2d2fd6c"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.078991 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-pkgzx" event={"ID":"65735728-e734-40de-a185-9caa1db9a47d","Type":"ContainerStarted","Data":"1fb571973e057c8519a5db05840734117d94ed09a1e521d57cf4a92fa88cbcfb"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.081666 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-w4bzp" event={"ID":"501e504c-7910-47cb-a181-812a0d72f2db","Type":"ContainerStarted","Data":"f8ab1b0def6ea5b1022acdb9c87659d2c1c20a66a736d1e630b89d097c36d6ef"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.081687 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-w4bzp" event={"ID":"501e504c-7910-47cb-a181-812a0d72f2db","Type":"ContainerStarted","Data":"bfa16e47d799d6c141954f2d1749fd776efec59e0c2322be83d41eba7af198db"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.090246 4948 generic.go:334] "Generic (PLEG): container finished" podID="10eedd01-3f52-4d44-981d-27bc16af186b" containerID="7e3cc540d00613674fe548789ab2ccc71a06c5714f420cedea97409a8254b14f" exitCode=0 Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.090309 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4djkr" event={"ID":"10eedd01-3f52-4d44-981d-27bc16af186b","Type":"ContainerDied","Data":"7e3cc540d00613674fe548789ab2ccc71a06c5714f420cedea97409a8254b14f"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.090337 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4djkr" event={"ID":"10eedd01-3f52-4d44-981d-27bc16af186b","Type":"ContainerStarted","Data":"f4a16b89f1881652510d924497279d30d07a12e1a23c9a2a09460395d54b6754"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.091556 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gz5j7" event={"ID":"b9370de8-8b72-4819-af79-4a9ee1fb6777","Type":"ContainerStarted","Data":"fd31b197071ac286c0663672b75ab91e334d6c66ca4e2bf75a86eb8ab8c0137c"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.093953 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-97j89" event={"ID":"4ced4ac5-e057-476a-908a-f573a8cf997b","Type":"ContainerStarted","Data":"fb257235a0241c571d44f0c21ce5fac86cd6393e8a1250cb5e78d4e1b599fb51"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.094042 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-97j89" event={"ID":"4ced4ac5-e057-476a-908a-f573a8cf997b","Type":"ContainerStarted","Data":"74d816dd0d8f2d728e06f6752ea62cf407348186231cd23345763d2a41ae00ad"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.102736 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-zvzvk" podStartSLOduration=3.102718112 podStartE2EDuration="3.102718112s" podCreationTimestamp="2026-02-20 08:21:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:58.09172735 +0000 UTC m=+967.066222170" watchObservedRunningTime="2026-02-20 08:21:58.102718112 +0000 UTC m=+967.077212932" Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.104118 4948 generic.go:334] "Generic (PLEG): container finished" podID="d3f86085-113f-408e-b43c-a499e2fe8b90" containerID="c565785e94684cef9e257c737e709e95f61b9c29c0e4f1d8e37358207acddb4c" exitCode=0 Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.104185 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a86b-account-create-update-mspp7" event={"ID":"d3f86085-113f-408e-b43c-a499e2fe8b90","Type":"ContainerDied","Data":"c565785e94684cef9e257c737e709e95f61b9c29c0e4f1d8e37358207acddb4c"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.104213 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a86b-account-create-update-mspp7" event={"ID":"d3f86085-113f-408e-b43c-a499e2fe8b90","Type":"ContainerStarted","Data":"b414d35cd5df6a0028614a71f6eb135810ed8113c4c0d12d90935b0df5cc95f0"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.111945 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0b2f-account-create-update-2f74q" event={"ID":"b6d068f3-210b-41e8-b90a-5406759c0606","Type":"ContainerStarted","Data":"a11f29a4f832ccc8dd04f52fcb05e1f418934455b1143a9d46f2d5ad1bce9425"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.112012 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0b2f-account-create-update-2f74q" event={"ID":"b6d068f3-210b-41e8-b90a-5406759c0606","Type":"ContainerStarted","Data":"67031a60eeb343261bf69ab0502142ef436d8c88a1348d9a5bdd6561d7700913"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.117121 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e6d0-account-create-update-cdxd5" event={"ID":"0a739e4d-8960-46f3-8c5d-2373b59165dd","Type":"ContainerStarted","Data":"ddcd74111702fbb6556b1046c04ec320ea3aecaa381a53da7a6aa1bc0c3d6fd7"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.117165 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e6d0-account-create-update-cdxd5" event={"ID":"0a739e4d-8960-46f3-8c5d-2373b59165dd","Type":"ContainerStarted","Data":"1e6ec887d9d3e16f02f5b4d5f7d7d830aca0b26b2e6976000ecd8e9ec7ae4ac5"} Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.118571 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-w4bzp" podStartSLOduration=4.118551864 podStartE2EDuration="4.118551864s" podCreationTimestamp="2026-02-20 08:21:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:58.115212081 +0000 UTC m=+967.089706901" watchObservedRunningTime="2026-02-20 08:21:58.118551864 +0000 UTC m=+967.093046684" Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.137248 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-pkgzx" podStartSLOduration=2.827735116 podStartE2EDuration="16.137230616s" podCreationTimestamp="2026-02-20 08:21:42 +0000 UTC" firstStartedPulling="2026-02-20 08:21:43.400951118 +0000 UTC m=+952.375445938" lastFinishedPulling="2026-02-20 08:21:56.710446618 +0000 UTC m=+965.684941438" observedRunningTime="2026-02-20 08:21:58.134117019 +0000 UTC m=+967.108611839" watchObservedRunningTime="2026-02-20 08:21:58.137230616 +0000 UTC m=+967.111725436" Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.184501 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-0b2f-account-create-update-2f74q" podStartSLOduration=4.184445105 podStartE2EDuration="4.184445105s" podCreationTimestamp="2026-02-20 08:21:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:58.183594904 +0000 UTC m=+967.158089724" watchObservedRunningTime="2026-02-20 08:21:58.184445105 +0000 UTC m=+967.158939925" Feb 20 08:21:58 crc kubenswrapper[4948]: I0220 08:21:58.184801 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-e6d0-account-create-update-cdxd5" podStartSLOduration=4.184793683 podStartE2EDuration="4.184793683s" podCreationTimestamp="2026-02-20 08:21:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:21:58.16810899 +0000 UTC m=+967.142603810" watchObservedRunningTime="2026-02-20 08:21:58.184793683 +0000 UTC m=+967.159288493" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.125098 4948 generic.go:334] "Generic (PLEG): container finished" podID="b6d068f3-210b-41e8-b90a-5406759c0606" containerID="a11f29a4f832ccc8dd04f52fcb05e1f418934455b1143a9d46f2d5ad1bce9425" exitCode=0 Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.125192 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0b2f-account-create-update-2f74q" event={"ID":"b6d068f3-210b-41e8-b90a-5406759c0606","Type":"ContainerDied","Data":"a11f29a4f832ccc8dd04f52fcb05e1f418934455b1143a9d46f2d5ad1bce9425"} Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.126952 4948 generic.go:334] "Generic (PLEG): container finished" podID="0a739e4d-8960-46f3-8c5d-2373b59165dd" containerID="ddcd74111702fbb6556b1046c04ec320ea3aecaa381a53da7a6aa1bc0c3d6fd7" exitCode=0 Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.127083 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e6d0-account-create-update-cdxd5" event={"ID":"0a739e4d-8960-46f3-8c5d-2373b59165dd","Type":"ContainerDied","Data":"ddcd74111702fbb6556b1046c04ec320ea3aecaa381a53da7a6aa1bc0c3d6fd7"} Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.128734 4948 generic.go:334] "Generic (PLEG): container finished" podID="501e504c-7910-47cb-a181-812a0d72f2db" containerID="f8ab1b0def6ea5b1022acdb9c87659d2c1c20a66a736d1e630b89d097c36d6ef" exitCode=0 Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.128795 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-w4bzp" event={"ID":"501e504c-7910-47cb-a181-812a0d72f2db","Type":"ContainerDied","Data":"f8ab1b0def6ea5b1022acdb9c87659d2c1c20a66a736d1e630b89d097c36d6ef"} Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.138737 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"0def77fdf4b8be6f4de61d06a4e0076ed0df40fa4e6037fcfcf3199187bc2f02"} Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.138772 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"7be95872e25f6acaeb57ea425561f0eff16c5602eb9dd465855ec55c34f78c37"} Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.138782 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"8b7432bd3983a8cca2ad3cbb32489a07f0374a62db1725e2a564b82efa942d38"} Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.154016 4948 generic.go:334] "Generic (PLEG): container finished" podID="4ced4ac5-e057-476a-908a-f573a8cf997b" containerID="fb257235a0241c571d44f0c21ce5fac86cd6393e8a1250cb5e78d4e1b599fb51" exitCode=0 Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.154061 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-97j89" event={"ID":"4ced4ac5-e057-476a-908a-f573a8cf997b","Type":"ContainerDied","Data":"fb257235a0241c571d44f0c21ce5fac86cd6393e8a1250cb5e78d4e1b599fb51"} Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.155810 4948 generic.go:334] "Generic (PLEG): container finished" podID="12f1f665-7a04-4794-a727-cc5c060a1933" containerID="cddaaa0c08e72fa702328c57550853b538c4537573467ac0ebd0761e1fcbc30c" exitCode=0 Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.155884 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-zvzvk" event={"ID":"12f1f665-7a04-4794-a727-cc5c060a1933","Type":"ContainerDied","Data":"cddaaa0c08e72fa702328c57550853b538c4537573467ac0ebd0761e1fcbc30c"} Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.562193 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.662270 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-97j89" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.668034 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4djkr" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.716276 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k984h\" (UniqueName: \"kubernetes.io/projected/d3f86085-113f-408e-b43c-a499e2fe8b90-kube-api-access-k984h\") pod \"d3f86085-113f-408e-b43c-a499e2fe8b90\" (UID: \"d3f86085-113f-408e-b43c-a499e2fe8b90\") " Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.716351 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3f86085-113f-408e-b43c-a499e2fe8b90-operator-scripts\") pod \"d3f86085-113f-408e-b43c-a499e2fe8b90\" (UID: \"d3f86085-113f-408e-b43c-a499e2fe8b90\") " Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.717414 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3f86085-113f-408e-b43c-a499e2fe8b90-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3f86085-113f-408e-b43c-a499e2fe8b90" (UID: "d3f86085-113f-408e-b43c-a499e2fe8b90"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.717906 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3f86085-113f-408e-b43c-a499e2fe8b90-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.721620 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3f86085-113f-408e-b43c-a499e2fe8b90-kube-api-access-k984h" (OuterVolumeSpecName: "kube-api-access-k984h") pod "d3f86085-113f-408e-b43c-a499e2fe8b90" (UID: "d3f86085-113f-408e-b43c-a499e2fe8b90"). InnerVolumeSpecName "kube-api-access-k984h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.818966 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrc6c\" (UniqueName: \"kubernetes.io/projected/4ced4ac5-e057-476a-908a-f573a8cf997b-kube-api-access-rrc6c\") pod \"4ced4ac5-e057-476a-908a-f573a8cf997b\" (UID: \"4ced4ac5-e057-476a-908a-f573a8cf997b\") " Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.819090 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kr2kf\" (UniqueName: \"kubernetes.io/projected/10eedd01-3f52-4d44-981d-27bc16af186b-kube-api-access-kr2kf\") pod \"10eedd01-3f52-4d44-981d-27bc16af186b\" (UID: \"10eedd01-3f52-4d44-981d-27bc16af186b\") " Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.819133 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ced4ac5-e057-476a-908a-f573a8cf997b-operator-scripts\") pod \"4ced4ac5-e057-476a-908a-f573a8cf997b\" (UID: \"4ced4ac5-e057-476a-908a-f573a8cf997b\") " Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.819319 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10eedd01-3f52-4d44-981d-27bc16af186b-operator-scripts\") pod \"10eedd01-3f52-4d44-981d-27bc16af186b\" (UID: \"10eedd01-3f52-4d44-981d-27bc16af186b\") " Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.819649 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ced4ac5-e057-476a-908a-f573a8cf997b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4ced4ac5-e057-476a-908a-f573a8cf997b" (UID: "4ced4ac5-e057-476a-908a-f573a8cf997b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.820671 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k984h\" (UniqueName: \"kubernetes.io/projected/d3f86085-113f-408e-b43c-a499e2fe8b90-kube-api-access-k984h\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.820720 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ced4ac5-e057-476a-908a-f573a8cf997b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.821103 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10eedd01-3f52-4d44-981d-27bc16af186b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "10eedd01-3f52-4d44-981d-27bc16af186b" (UID: "10eedd01-3f52-4d44-981d-27bc16af186b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.823267 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10eedd01-3f52-4d44-981d-27bc16af186b-kube-api-access-kr2kf" (OuterVolumeSpecName: "kube-api-access-kr2kf") pod "10eedd01-3f52-4d44-981d-27bc16af186b" (UID: "10eedd01-3f52-4d44-981d-27bc16af186b"). InnerVolumeSpecName "kube-api-access-kr2kf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.824770 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ced4ac5-e057-476a-908a-f573a8cf997b-kube-api-access-rrc6c" (OuterVolumeSpecName: "kube-api-access-rrc6c") pod "4ced4ac5-e057-476a-908a-f573a8cf997b" (UID: "4ced4ac5-e057-476a-908a-f573a8cf997b"). InnerVolumeSpecName "kube-api-access-rrc6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.922951 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10eedd01-3f52-4d44-981d-27bc16af186b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.923018 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrc6c\" (UniqueName: \"kubernetes.io/projected/4ced4ac5-e057-476a-908a-f573a8cf997b-kube-api-access-rrc6c\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:59 crc kubenswrapper[4948]: I0220 08:21:59.923035 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kr2kf\" (UniqueName: \"kubernetes.io/projected/10eedd01-3f52-4d44-981d-27bc16af186b-kube-api-access-kr2kf\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.165014 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-97j89" event={"ID":"4ced4ac5-e057-476a-908a-f573a8cf997b","Type":"ContainerDied","Data":"74d816dd0d8f2d728e06f6752ea62cf407348186231cd23345763d2a41ae00ad"} Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.166220 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74d816dd0d8f2d728e06f6752ea62cf407348186231cd23345763d2a41ae00ad" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.166425 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-97j89" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.192381 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a86b-account-create-update-mspp7" event={"ID":"d3f86085-113f-408e-b43c-a499e2fe8b90","Type":"ContainerDied","Data":"b414d35cd5df6a0028614a71f6eb135810ed8113c4c0d12d90935b0df5cc95f0"} Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.192443 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b414d35cd5df6a0028614a71f6eb135810ed8113c4c0d12d90935b0df5cc95f0" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.192405 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a86b-account-create-update-mspp7" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.199775 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4djkr" event={"ID":"10eedd01-3f52-4d44-981d-27bc16af186b","Type":"ContainerDied","Data":"f4a16b89f1881652510d924497279d30d07a12e1a23c9a2a09460395d54b6754"} Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.199826 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4a16b89f1881652510d924497279d30d07a12e1a23c9a2a09460395d54b6754" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.200033 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4djkr" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.205354 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"f0e914fe99d8fbabaf9e4af16ab7bd5f4b91015e66934992a5c19a45d77421d7"} Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.634165 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.741183 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dqq5\" (UniqueName: \"kubernetes.io/projected/0a739e4d-8960-46f3-8c5d-2373b59165dd-kube-api-access-2dqq5\") pod \"0a739e4d-8960-46f3-8c5d-2373b59165dd\" (UID: \"0a739e4d-8960-46f3-8c5d-2373b59165dd\") " Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.741461 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a739e4d-8960-46f3-8c5d-2373b59165dd-operator-scripts\") pod \"0a739e4d-8960-46f3-8c5d-2373b59165dd\" (UID: \"0a739e4d-8960-46f3-8c5d-2373b59165dd\") " Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.742899 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a739e4d-8960-46f3-8c5d-2373b59165dd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0a739e4d-8960-46f3-8c5d-2373b59165dd" (UID: "0a739e4d-8960-46f3-8c5d-2373b59165dd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.766236 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a739e4d-8960-46f3-8c5d-2373b59165dd-kube-api-access-2dqq5" (OuterVolumeSpecName: "kube-api-access-2dqq5") pod "0a739e4d-8960-46f3-8c5d-2373b59165dd" (UID: "0a739e4d-8960-46f3-8c5d-2373b59165dd"). InnerVolumeSpecName "kube-api-access-2dqq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.843453 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dqq5\" (UniqueName: \"kubernetes.io/projected/0a739e4d-8960-46f3-8c5d-2373b59165dd-kube-api-access-2dqq5\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.843486 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a739e4d-8960-46f3-8c5d-2373b59165dd-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.886402 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-w4bzp" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.894892 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:22:00 crc kubenswrapper[4948]: I0220 08:22:00.900158 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-zvzvk" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.046596 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6d068f3-210b-41e8-b90a-5406759c0606-operator-scripts\") pod \"b6d068f3-210b-41e8-b90a-5406759c0606\" (UID: \"b6d068f3-210b-41e8-b90a-5406759c0606\") " Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.046685 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12f1f665-7a04-4794-a727-cc5c060a1933-operator-scripts\") pod \"12f1f665-7a04-4794-a727-cc5c060a1933\" (UID: \"12f1f665-7a04-4794-a727-cc5c060a1933\") " Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.046726 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/501e504c-7910-47cb-a181-812a0d72f2db-operator-scripts\") pod \"501e504c-7910-47cb-a181-812a0d72f2db\" (UID: \"501e504c-7910-47cb-a181-812a0d72f2db\") " Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.046793 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcv9f\" (UniqueName: \"kubernetes.io/projected/12f1f665-7a04-4794-a727-cc5c060a1933-kube-api-access-rcv9f\") pod \"12f1f665-7a04-4794-a727-cc5c060a1933\" (UID: \"12f1f665-7a04-4794-a727-cc5c060a1933\") " Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.046860 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvqt6\" (UniqueName: \"kubernetes.io/projected/501e504c-7910-47cb-a181-812a0d72f2db-kube-api-access-bvqt6\") pod \"501e504c-7910-47cb-a181-812a0d72f2db\" (UID: \"501e504c-7910-47cb-a181-812a0d72f2db\") " Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.046883 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c86dh\" (UniqueName: \"kubernetes.io/projected/b6d068f3-210b-41e8-b90a-5406759c0606-kube-api-access-c86dh\") pod \"b6d068f3-210b-41e8-b90a-5406759c0606\" (UID: \"b6d068f3-210b-41e8-b90a-5406759c0606\") " Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.047611 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/501e504c-7910-47cb-a181-812a0d72f2db-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "501e504c-7910-47cb-a181-812a0d72f2db" (UID: "501e504c-7910-47cb-a181-812a0d72f2db"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.047696 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6d068f3-210b-41e8-b90a-5406759c0606-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b6d068f3-210b-41e8-b90a-5406759c0606" (UID: "b6d068f3-210b-41e8-b90a-5406759c0606"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.047917 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12f1f665-7a04-4794-a727-cc5c060a1933-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "12f1f665-7a04-4794-a727-cc5c060a1933" (UID: "12f1f665-7a04-4794-a727-cc5c060a1933"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.053202 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6d068f3-210b-41e8-b90a-5406759c0606-kube-api-access-c86dh" (OuterVolumeSpecName: "kube-api-access-c86dh") pod "b6d068f3-210b-41e8-b90a-5406759c0606" (UID: "b6d068f3-210b-41e8-b90a-5406759c0606"). InnerVolumeSpecName "kube-api-access-c86dh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.053295 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12f1f665-7a04-4794-a727-cc5c060a1933-kube-api-access-rcv9f" (OuterVolumeSpecName: "kube-api-access-rcv9f") pod "12f1f665-7a04-4794-a727-cc5c060a1933" (UID: "12f1f665-7a04-4794-a727-cc5c060a1933"). InnerVolumeSpecName "kube-api-access-rcv9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.058950 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/501e504c-7910-47cb-a181-812a0d72f2db-kube-api-access-bvqt6" (OuterVolumeSpecName: "kube-api-access-bvqt6") pod "501e504c-7910-47cb-a181-812a0d72f2db" (UID: "501e504c-7910-47cb-a181-812a0d72f2db"). InnerVolumeSpecName "kube-api-access-bvqt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.149106 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12f1f665-7a04-4794-a727-cc5c060a1933-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.149142 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/501e504c-7910-47cb-a181-812a0d72f2db-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.149151 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcv9f\" (UniqueName: \"kubernetes.io/projected/12f1f665-7a04-4794-a727-cc5c060a1933-kube-api-access-rcv9f\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.149163 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvqt6\" (UniqueName: \"kubernetes.io/projected/501e504c-7910-47cb-a181-812a0d72f2db-kube-api-access-bvqt6\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.149172 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c86dh\" (UniqueName: \"kubernetes.io/projected/b6d068f3-210b-41e8-b90a-5406759c0606-kube-api-access-c86dh\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.149180 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6d068f3-210b-41e8-b90a-5406759c0606-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.228186 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"3355e23b27ec2e8d34fef5fb30fe085eb1135b951828ca9cccf056bb4392c599"} Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.228230 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"73d89eec04dd76cb9501192d2204144dca67ce3fa8d711ecff913828c149145f"} Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.233945 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-zvzvk" event={"ID":"12f1f665-7a04-4794-a727-cc5c060a1933","Type":"ContainerDied","Data":"4ff5b9d3c7487921f2af53697fe4d63849f3213b42b4935658e6dfe3c2d2fd6c"} Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.234033 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ff5b9d3c7487921f2af53697fe4d63849f3213b42b4935658e6dfe3c2d2fd6c" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.233958 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-zvzvk" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.236261 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e6d0-account-create-update-cdxd5" event={"ID":"0a739e4d-8960-46f3-8c5d-2373b59165dd","Type":"ContainerDied","Data":"1e6ec887d9d3e16f02f5b4d5f7d7d830aca0b26b2e6976000ecd8e9ec7ae4ac5"} Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.236294 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e6ec887d9d3e16f02f5b4d5f7d7d830aca0b26b2e6976000ecd8e9ec7ae4ac5" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.236360 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e6d0-account-create-update-cdxd5" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.241933 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0b2f-account-create-update-2f74q" event={"ID":"b6d068f3-210b-41e8-b90a-5406759c0606","Type":"ContainerDied","Data":"67031a60eeb343261bf69ab0502142ef436d8c88a1348d9a5bdd6561d7700913"} Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.241991 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67031a60eeb343261bf69ab0502142ef436d8c88a1348d9a5bdd6561d7700913" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.242046 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0b2f-account-create-update-2f74q" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.247286 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-w4bzp" event={"ID":"501e504c-7910-47cb-a181-812a0d72f2db","Type":"ContainerDied","Data":"bfa16e47d799d6c141954f2d1749fd776efec59e0c2322be83d41eba7af198db"} Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.247321 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfa16e47d799d6c141954f2d1749fd776efec59e0c2322be83d41eba7af198db" Feb 20 08:22:01 crc kubenswrapper[4948]: I0220 08:22:01.247370 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-w4bzp" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.444547 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lnp7g"] Feb 20 08:22:03 crc kubenswrapper[4948]: E0220 08:22:03.445311 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12f1f665-7a04-4794-a727-cc5c060a1933" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445329 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="12f1f665-7a04-4794-a727-cc5c060a1933" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: E0220 08:22:03.445347 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ced4ac5-e057-476a-908a-f573a8cf997b" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445357 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ced4ac5-e057-476a-908a-f573a8cf997b" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: E0220 08:22:03.445370 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501e504c-7910-47cb-a181-812a0d72f2db" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445378 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="501e504c-7910-47cb-a181-812a0d72f2db" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: E0220 08:22:03.445396 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d068f3-210b-41e8-b90a-5406759c0606" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445403 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d068f3-210b-41e8-b90a-5406759c0606" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: E0220 08:22:03.445419 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10eedd01-3f52-4d44-981d-27bc16af186b" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445427 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="10eedd01-3f52-4d44-981d-27bc16af186b" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: E0220 08:22:03.445446 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f86085-113f-408e-b43c-a499e2fe8b90" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445455 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f86085-113f-408e-b43c-a499e2fe8b90" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: E0220 08:22:03.445464 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa672fa0-b98b-477c-a34c-ce996ce6592d" containerName="ovn-config" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445472 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa672fa0-b98b-477c-a34c-ce996ce6592d" containerName="ovn-config" Feb 20 08:22:03 crc kubenswrapper[4948]: E0220 08:22:03.445491 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a739e4d-8960-46f3-8c5d-2373b59165dd" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445499 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a739e4d-8960-46f3-8c5d-2373b59165dd" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445729 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="10eedd01-3f52-4d44-981d-27bc16af186b" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445747 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a739e4d-8960-46f3-8c5d-2373b59165dd" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445754 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="501e504c-7910-47cb-a181-812a0d72f2db" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445765 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d068f3-210b-41e8-b90a-5406759c0606" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445782 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa672fa0-b98b-477c-a34c-ce996ce6592d" containerName="ovn-config" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445791 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f86085-113f-408e-b43c-a499e2fe8b90" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445802 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="12f1f665-7a04-4794-a727-cc5c060a1933" containerName="mariadb-account-create-update" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.445811 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ced4ac5-e057-476a-908a-f573a8cf997b" containerName="mariadb-database-create" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.454474 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.495055 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnp7g"] Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.585080 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-utilities\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.585119 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-catalog-content\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.585144 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5vn9\" (UniqueName: \"kubernetes.io/projected/81e7d5de-2379-46c7-ba9b-ffcacad79e29-kube-api-access-w5vn9\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.686906 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-utilities\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.686938 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-catalog-content\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.686959 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5vn9\" (UniqueName: \"kubernetes.io/projected/81e7d5de-2379-46c7-ba9b-ffcacad79e29-kube-api-access-w5vn9\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.687740 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-catalog-content\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.687835 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-utilities\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.716533 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5vn9\" (UniqueName: \"kubernetes.io/projected/81e7d5de-2379-46c7-ba9b-ffcacad79e29-kube-api-access-w5vn9\") pod \"redhat-marketplace-lnp7g\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:03 crc kubenswrapper[4948]: I0220 08:22:03.796416 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:04 crc kubenswrapper[4948]: I0220 08:22:04.256875 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnp7g"] Feb 20 08:22:04 crc kubenswrapper[4948]: W0220 08:22:04.266236 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81e7d5de_2379_46c7_ba9b_ffcacad79e29.slice/crio-ac4ceb2acdc56fbdc4b52d619a840394196af858618bbcf55bc7c1dc060d1a2f WatchSource:0}: Error finding container ac4ceb2acdc56fbdc4b52d619a840394196af858618bbcf55bc7c1dc060d1a2f: Status 404 returned error can't find the container with id ac4ceb2acdc56fbdc4b52d619a840394196af858618bbcf55bc7c1dc060d1a2f Feb 20 08:22:04 crc kubenswrapper[4948]: I0220 08:22:04.278565 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"5369372614d8e43350703b61475cc193540f62c9e2a5e484b1fbcd3c24f8cf51"} Feb 20 08:22:04 crc kubenswrapper[4948]: I0220 08:22:04.278944 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"6c3625774a56cd70321fcb0287393d4cc54f0526941e165f455cd535741f10cd"} Feb 20 08:22:04 crc kubenswrapper[4948]: I0220 08:22:04.280624 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gz5j7" event={"ID":"b9370de8-8b72-4819-af79-4a9ee1fb6777","Type":"ContainerStarted","Data":"72b50822f4c48e3166b0066730c258d864d6094e98d9d9617ca0abdc1f36ae38"} Feb 20 08:22:04 crc kubenswrapper[4948]: I0220 08:22:04.307272 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-gz5j7" podStartSLOduration=4.165632658 podStartE2EDuration="10.307252965s" podCreationTimestamp="2026-02-20 08:21:54 +0000 UTC" firstStartedPulling="2026-02-20 08:21:57.624738803 +0000 UTC m=+966.599233623" lastFinishedPulling="2026-02-20 08:22:03.76635911 +0000 UTC m=+972.740853930" observedRunningTime="2026-02-20 08:22:04.303165644 +0000 UTC m=+973.277660464" watchObservedRunningTime="2026-02-20 08:22:04.307252965 +0000 UTC m=+973.281747785" Feb 20 08:22:05 crc kubenswrapper[4948]: I0220 08:22:05.299865 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"f5d776784274a96fef8898bd072052c1535eba851190d0eacd8ac02fe6ab20f6"} Feb 20 08:22:05 crc kubenswrapper[4948]: I0220 08:22:05.300017 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"e92a5dd37239b7f9de68a6175ce427a7d03ddfec618577a2b42564193f86893a"} Feb 20 08:22:05 crc kubenswrapper[4948]: I0220 08:22:05.301880 4948 generic.go:334] "Generic (PLEG): container finished" podID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerID="82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b" exitCode=0 Feb 20 08:22:05 crc kubenswrapper[4948]: I0220 08:22:05.302067 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnp7g" event={"ID":"81e7d5de-2379-46c7-ba9b-ffcacad79e29","Type":"ContainerDied","Data":"82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b"} Feb 20 08:22:05 crc kubenswrapper[4948]: I0220 08:22:05.302096 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnp7g" event={"ID":"81e7d5de-2379-46c7-ba9b-ffcacad79e29","Type":"ContainerStarted","Data":"ac4ceb2acdc56fbdc4b52d619a840394196af858618bbcf55bc7c1dc060d1a2f"} Feb 20 08:22:09 crc kubenswrapper[4948]: I0220 08:22:09.340378 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e4d60def-5dc9-460b-92ed-e3d0157982ec","Type":"ContainerStarted","Data":"38118f7052314e9310bad19f8bc505ccd82a8634fb080d44e8b173fffa583cd6"} Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.349420 4948 generic.go:334] "Generic (PLEG): container finished" podID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerID="833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15" exitCode=0 Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.349469 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnp7g" event={"ID":"81e7d5de-2379-46c7-ba9b-ffcacad79e29","Type":"ContainerDied","Data":"833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15"} Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.352601 4948 generic.go:334] "Generic (PLEG): container finished" podID="65735728-e734-40de-a185-9caa1db9a47d" containerID="1fb571973e057c8519a5db05840734117d94ed09a1e521d57cf4a92fa88cbcfb" exitCode=0 Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.352687 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-pkgzx" event={"ID":"65735728-e734-40de-a185-9caa1db9a47d","Type":"ContainerDied","Data":"1fb571973e057c8519a5db05840734117d94ed09a1e521d57cf4a92fa88cbcfb"} Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.405108 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=27.962875945 podStartE2EDuration="41.405082558s" podCreationTimestamp="2026-02-20 08:21:29 +0000 UTC" firstStartedPulling="2026-02-20 08:21:47.117874941 +0000 UTC m=+956.092369761" lastFinishedPulling="2026-02-20 08:22:00.560081544 +0000 UTC m=+969.534576374" observedRunningTime="2026-02-20 08:22:10.399139931 +0000 UTC m=+979.373634761" watchObservedRunningTime="2026-02-20 08:22:10.405082558 +0000 UTC m=+979.379577398" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.698203 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-mvrcc"] Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.699760 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.708181 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.716267 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-mvrcc"] Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.798987 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-svc\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.799340 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.799405 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-config\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.799436 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.799497 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.799565 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57xf5\" (UniqueName: \"kubernetes.io/projected/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-kube-api-access-57xf5\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.900546 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.900607 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57xf5\" (UniqueName: \"kubernetes.io/projected/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-kube-api-access-57xf5\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.900652 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-svc\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.900668 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.900692 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-config\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.900718 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.901619 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.901624 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-svc\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.901775 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.901827 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-config\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.903494 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:10 crc kubenswrapper[4948]: I0220 08:22:10.922361 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57xf5\" (UniqueName: \"kubernetes.io/projected/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-kube-api-access-57xf5\") pod \"dnsmasq-dns-764c5664d7-mvrcc\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.024230 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.362144 4948 generic.go:334] "Generic (PLEG): container finished" podID="b9370de8-8b72-4819-af79-4a9ee1fb6777" containerID="72b50822f4c48e3166b0066730c258d864d6094e98d9d9617ca0abdc1f36ae38" exitCode=0 Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.362222 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gz5j7" event={"ID":"b9370de8-8b72-4819-af79-4a9ee1fb6777","Type":"ContainerDied","Data":"72b50822f4c48e3166b0066730c258d864d6094e98d9d9617ca0abdc1f36ae38"} Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.364563 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnp7g" event={"ID":"81e7d5de-2379-46c7-ba9b-ffcacad79e29","Type":"ContainerStarted","Data":"59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68"} Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.402195 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lnp7g" podStartSLOduration=2.957574144 podStartE2EDuration="8.402179142s" podCreationTimestamp="2026-02-20 08:22:03 +0000 UTC" firstStartedPulling="2026-02-20 08:22:05.303570081 +0000 UTC m=+974.278064901" lastFinishedPulling="2026-02-20 08:22:10.748175079 +0000 UTC m=+979.722669899" observedRunningTime="2026-02-20 08:22:11.39563659 +0000 UTC m=+980.370131410" watchObservedRunningTime="2026-02-20 08:22:11.402179142 +0000 UTC m=+980.376673962" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.548678 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-mvrcc"] Feb 20 08:22:11 crc kubenswrapper[4948]: W0220 08:22:11.551211 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b8788ed_d6f3_4344_b8b2_7bb554ecd25e.slice/crio-5b5cca04b9c3f166250baa067797237d67a3b5df1ae92815a7da886a955a6658 WatchSource:0}: Error finding container 5b5cca04b9c3f166250baa067797237d67a3b5df1ae92815a7da886a955a6658: Status 404 returned error can't find the container with id 5b5cca04b9c3f166250baa067797237d67a3b5df1ae92815a7da886a955a6658 Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.714729 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-pkgzx" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.828440 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-config-data\") pod \"65735728-e734-40de-a185-9caa1db9a47d\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.828578 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j589d\" (UniqueName: \"kubernetes.io/projected/65735728-e734-40de-a185-9caa1db9a47d-kube-api-access-j589d\") pod \"65735728-e734-40de-a185-9caa1db9a47d\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.828602 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-db-sync-config-data\") pod \"65735728-e734-40de-a185-9caa1db9a47d\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.828681 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-combined-ca-bundle\") pod \"65735728-e734-40de-a185-9caa1db9a47d\" (UID: \"65735728-e734-40de-a185-9caa1db9a47d\") " Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.833647 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "65735728-e734-40de-a185-9caa1db9a47d" (UID: "65735728-e734-40de-a185-9caa1db9a47d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.834007 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65735728-e734-40de-a185-9caa1db9a47d-kube-api-access-j589d" (OuterVolumeSpecName: "kube-api-access-j589d") pod "65735728-e734-40de-a185-9caa1db9a47d" (UID: "65735728-e734-40de-a185-9caa1db9a47d"). InnerVolumeSpecName "kube-api-access-j589d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.850375 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65735728-e734-40de-a185-9caa1db9a47d" (UID: "65735728-e734-40de-a185-9caa1db9a47d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.869861 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-config-data" (OuterVolumeSpecName: "config-data") pod "65735728-e734-40de-a185-9caa1db9a47d" (UID: "65735728-e734-40de-a185-9caa1db9a47d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.931280 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j589d\" (UniqueName: \"kubernetes.io/projected/65735728-e734-40de-a185-9caa1db9a47d-kube-api-access-j589d\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.931336 4948 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.931357 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:11 crc kubenswrapper[4948]: I0220 08:22:11.931374 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65735728-e734-40de-a185-9caa1db9a47d-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.375771 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-pkgzx" event={"ID":"65735728-e734-40de-a185-9caa1db9a47d","Type":"ContainerDied","Data":"5e99cbf3924927a30ad54b19ba1a0b9e160811b4e81053363ade5b0662fa69fa"} Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.375821 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e99cbf3924927a30ad54b19ba1a0b9e160811b4e81053363ade5b0662fa69fa" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.375912 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-pkgzx" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.378705 4948 generic.go:334] "Generic (PLEG): container finished" podID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" containerID="596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58" exitCode=0 Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.378753 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" event={"ID":"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e","Type":"ContainerDied","Data":"596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58"} Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.378921 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" event={"ID":"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e","Type":"ContainerStarted","Data":"5b5cca04b9c3f166250baa067797237d67a3b5df1ae92815a7da886a955a6658"} Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.763878 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-mvrcc"] Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.769891 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.811455 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-v6srk"] Feb 20 08:22:12 crc kubenswrapper[4948]: E0220 08:22:12.811927 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65735728-e734-40de-a185-9caa1db9a47d" containerName="glance-db-sync" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.811954 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="65735728-e734-40de-a185-9caa1db9a47d" containerName="glance-db-sync" Feb 20 08:22:12 crc kubenswrapper[4948]: E0220 08:22:12.811967 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9370de8-8b72-4819-af79-4a9ee1fb6777" containerName="keystone-db-sync" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.812027 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9370de8-8b72-4819-af79-4a9ee1fb6777" containerName="keystone-db-sync" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.812231 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="65735728-e734-40de-a185-9caa1db9a47d" containerName="glance-db-sync" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.812264 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9370de8-8b72-4819-af79-4a9ee1fb6777" containerName="keystone-db-sync" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.813334 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.822033 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-v6srk"] Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.847724 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-config-data\") pod \"b9370de8-8b72-4819-af79-4a9ee1fb6777\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.847866 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx8w6\" (UniqueName: \"kubernetes.io/projected/b9370de8-8b72-4819-af79-4a9ee1fb6777-kube-api-access-zx8w6\") pod \"b9370de8-8b72-4819-af79-4a9ee1fb6777\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.847921 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-combined-ca-bundle\") pod \"b9370de8-8b72-4819-af79-4a9ee1fb6777\" (UID: \"b9370de8-8b72-4819-af79-4a9ee1fb6777\") " Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.852693 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9370de8-8b72-4819-af79-4a9ee1fb6777-kube-api-access-zx8w6" (OuterVolumeSpecName: "kube-api-access-zx8w6") pod "b9370de8-8b72-4819-af79-4a9ee1fb6777" (UID: "b9370de8-8b72-4819-af79-4a9ee1fb6777"). InnerVolumeSpecName "kube-api-access-zx8w6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.875346 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b9370de8-8b72-4819-af79-4a9ee1fb6777" (UID: "b9370de8-8b72-4819-af79-4a9ee1fb6777"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.891446 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-config-data" (OuterVolumeSpecName: "config-data") pod "b9370de8-8b72-4819-af79-4a9ee1fb6777" (UID: "b9370de8-8b72-4819-af79-4a9ee1fb6777"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.950408 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.950501 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.950651 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74rwq\" (UniqueName: \"kubernetes.io/projected/7960b704-21ac-4d75-8226-a9bbae4c2b50-kube-api-access-74rwq\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.950729 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-config\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.950833 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.950887 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.951089 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.951101 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9370de8-8b72-4819-af79-4a9ee1fb6777-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:12 crc kubenswrapper[4948]: I0220 08:22:12.951111 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx8w6\" (UniqueName: \"kubernetes.io/projected/b9370de8-8b72-4819-af79-4a9ee1fb6777-kube-api-access-zx8w6\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.052360 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.052415 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.052494 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.052516 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.052553 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74rwq\" (UniqueName: \"kubernetes.io/projected/7960b704-21ac-4d75-8226-a9bbae4c2b50-kube-api-access-74rwq\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.052590 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-config\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.053512 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.054032 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.054049 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.054098 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-config\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.054360 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.085902 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74rwq\" (UniqueName: \"kubernetes.io/projected/7960b704-21ac-4d75-8226-a9bbae4c2b50-kube-api-access-74rwq\") pod \"dnsmasq-dns-74f6bcbc87-v6srk\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.232120 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.387931 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gz5j7" event={"ID":"b9370de8-8b72-4819-af79-4a9ee1fb6777","Type":"ContainerDied","Data":"fd31b197071ac286c0663672b75ab91e334d6c66ca4e2bf75a86eb8ab8c0137c"} Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.388013 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gz5j7" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.388032 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd31b197071ac286c0663672b75ab91e334d6c66ca4e2bf75a86eb8ab8c0137c" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.390382 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" event={"ID":"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e","Type":"ContainerStarted","Data":"cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d"} Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.391417 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.430499 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" podStartSLOduration=3.430481457 podStartE2EDuration="3.430481457s" podCreationTimestamp="2026-02-20 08:22:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:13.42453272 +0000 UTC m=+982.399027550" watchObservedRunningTime="2026-02-20 08:22:13.430481457 +0000 UTC m=+982.404976277" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.686129 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-v6srk"] Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.699164 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-g72bc"] Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.700422 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.704804 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.706884 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.707096 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.707287 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.707396 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-v9rj2" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.734509 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-g72bc"] Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.765941 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-fernet-keys\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.766028 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-combined-ca-bundle\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.766104 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-scripts\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.766130 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-credential-keys\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.766212 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-config-data\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.766235 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p27c6\" (UniqueName: \"kubernetes.io/projected/9959c3d4-fb0e-473c-8bbe-065110a3e03a-kube-api-access-p27c6\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.797202 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.800211 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.855084 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-gmdvb"] Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.862325 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.871848 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-scripts\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.871892 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-credential-keys\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.871987 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-config-data\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.872002 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p27c6\" (UniqueName: \"kubernetes.io/projected/9959c3d4-fb0e-473c-8bbe-065110a3e03a-kube-api-access-p27c6\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.872059 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-fernet-keys\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.872083 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-combined-ca-bundle\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.895572 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-credential-keys\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.904628 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-config-data\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.909268 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-fernet-keys\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.912447 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-combined-ca-bundle\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.925028 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-gmdvb"] Feb 20 08:22:13 crc kubenswrapper[4948]: W0220 08:22:13.925927 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7960b704_21ac_4d75_8226_a9bbae4c2b50.slice/crio-e3cb32ae5e929563597aaafd43d3401d85066f36163aa0afc9720c6e975f05be WatchSource:0}: Error finding container e3cb32ae5e929563597aaafd43d3401d85066f36163aa0afc9720c6e975f05be: Status 404 returned error can't find the container with id e3cb32ae5e929563597aaafd43d3401d85066f36163aa0afc9720c6e975f05be Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.942320 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-scripts\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.944902 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-v6srk"] Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.972352 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p27c6\" (UniqueName: \"kubernetes.io/projected/9959c3d4-fb0e-473c-8bbe-065110a3e03a-kube-api-access-p27c6\") pod \"keystone-bootstrap-g72bc\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.973960 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-config\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.974083 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.974145 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.974165 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-svc\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.974213 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28ln7\" (UniqueName: \"kubernetes.io/projected/69c38b92-a051-4834-bf34-8612edb31dfa-kube-api-access-28ln7\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.974229 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:13 crc kubenswrapper[4948]: I0220 08:22:13.979257 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.018862 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-g5zxh"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.019883 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.057006 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d578c747f-8f6hc"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.058348 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.066760 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.066944 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.067095 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4txrz" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.067270 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.067482 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.067620 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-75xnq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.067846 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.068266 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.081833 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.081887 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-svc\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.081956 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-combined-ca-bundle\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.082038 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-config\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.082265 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28ln7\" (UniqueName: \"kubernetes.io/projected/69c38b92-a051-4834-bf34-8612edb31dfa-kube-api-access-28ln7\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.082291 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.082379 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npm75\" (UniqueName: \"kubernetes.io/projected/bb362d0e-68b4-4d48-885c-a6894ac71e6e-kube-api-access-npm75\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.082437 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-config\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.082456 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.083225 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-svc\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.083795 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.094766 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.094769 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.095469 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-config\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.099423 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-g5zxh"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.166074 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d578c747f-8f6hc"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.192344 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-scripts\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.192675 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npm75\" (UniqueName: \"kubernetes.io/projected/bb362d0e-68b4-4d48-885c-a6894ac71e6e-kube-api-access-npm75\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.192695 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-logs\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.192735 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrsxt\" (UniqueName: \"kubernetes.io/projected/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-kube-api-access-jrsxt\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.192780 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-horizon-secret-key\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.192857 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-combined-ca-bundle\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.192872 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-config\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.192938 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-config-data\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.202244 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-combined-ca-bundle\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.207627 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28ln7\" (UniqueName: \"kubernetes.io/projected/69c38b92-a051-4834-bf34-8612edb31dfa-kube-api-access-28ln7\") pod \"dnsmasq-dns-847c4cc679-gmdvb\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.210083 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-xt2zq"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.226875 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-config\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.231166 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.241292 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-q6pbd" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.241475 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.246215 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.255949 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-xt2zq"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.265517 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npm75\" (UniqueName: \"kubernetes.io/projected/bb362d0e-68b4-4d48-885c-a6894ac71e6e-kube-api-access-npm75\") pod \"neutron-db-sync-g5zxh\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.268036 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.270074 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.280477 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.280702 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.294894 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f706b741-86f9-49a3-95d5-85ee62eb3668-etc-machine-id\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.294937 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-config-data\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.294997 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-scripts\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.295017 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-logs\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.295044 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrsxt\" (UniqueName: \"kubernetes.io/projected/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-kube-api-access-jrsxt\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.295075 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-horizon-secret-key\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.295109 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-config-data\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.295132 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-db-sync-config-data\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.295153 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-combined-ca-bundle\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.295168 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-scripts\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.295231 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwtv5\" (UniqueName: \"kubernetes.io/projected/f706b741-86f9-49a3-95d5-85ee62eb3668-kube-api-access-fwtv5\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.296342 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-config-data\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.300862 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.302036 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-logs\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.310350 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-scripts\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.330463 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-horizon-secret-key\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.363921 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrsxt\" (UniqueName: \"kubernetes.io/projected/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-kube-api-access-jrsxt\") pod \"horizon-7d578c747f-8f6hc\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.387032 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.388300 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402011 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-config-data\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402090 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402134 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-config-data\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402166 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-db-sync-config-data\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402187 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-combined-ca-bundle\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402208 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-scripts\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402262 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwtv5\" (UniqueName: \"kubernetes.io/projected/f706b741-86f9-49a3-95d5-85ee62eb3668-kube-api-access-fwtv5\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402299 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f706b741-86f9-49a3-95d5-85ee62eb3668-etc-machine-id\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402331 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402365 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-scripts\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402383 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77njn\" (UniqueName: \"kubernetes.io/projected/89b49c50-bae7-47fe-b85f-7a718953933d-kube-api-access-77njn\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402399 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-run-httpd\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.402423 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-log-httpd\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.403329 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-6mz7x"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.404761 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.405791 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.408093 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f706b741-86f9-49a3-95d5-85ee62eb3668-etc-machine-id\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.408601 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bgxj8" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.409149 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.417260 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.423073 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-gmdvb"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.432772 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-2fq52" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.432995 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.433105 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.433566 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-db-sync-config-data\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.434117 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-config-data\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.437605 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-combined-ca-bundle\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.440947 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-scripts\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.473391 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.478400 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwtv5\" (UniqueName: \"kubernetes.io/projected/f706b741-86f9-49a3-95d5-85ee62eb3668-kube-api-access-fwtv5\") pod \"cinder-db-sync-xt2zq\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.478476 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.488778 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.525120 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6mz7x"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.529552 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.529674 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1227b752-8a3c-4c01-91bf-0662c1b1a231-logs\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.529763 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.529802 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.529872 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-config-data\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.529984 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-logs\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530018 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530062 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rtp8\" (UniqueName: \"kubernetes.io/projected/1227b752-8a3c-4c01-91bf-0662c1b1a231-kube-api-access-5rtp8\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530094 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqhps\" (UniqueName: \"kubernetes.io/projected/b8ad7a10-6353-45f2-a11d-126d0d4786b0-kube-api-access-kqhps\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530126 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-scripts\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530149 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77njn\" (UniqueName: \"kubernetes.io/projected/89b49c50-bae7-47fe-b85f-7a718953933d-kube-api-access-77njn\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530173 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-run-httpd\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530214 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-scripts\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530238 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530268 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-log-httpd\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530316 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-config-data\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530358 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-combined-ca-bundle\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530391 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-config-data\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.530415 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-scripts\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.531703 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-run-httpd\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.542489 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-scripts\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.544520 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" event={"ID":"7960b704-21ac-4d75-8226-a9bbae4c2b50","Type":"ContainerStarted","Data":"e3cb32ae5e929563597aaafd43d3401d85066f36163aa0afc9720c6e975f05be"} Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.544823 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" podUID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" containerName="dnsmasq-dns" containerID="cri-o://cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d" gracePeriod=10 Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.548157 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-log-httpd\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.551150 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-config-data\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.554903 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77njn\" (UniqueName: \"kubernetes.io/projected/89b49c50-bae7-47fe-b85f-7a718953933d-kube-api-access-77njn\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.560201 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.563509 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.564114 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-c7zzv"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.565813 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.567984 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8rh2b" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.568152 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.573235 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-54d5f75fb9-s2659"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.574543 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.583966 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.588060 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-c7zzv"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.602375 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-96668"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.604308 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.605027 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.614329 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-54d5f75fb9-s2659"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.621264 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-96668"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636446 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1227b752-8a3c-4c01-91bf-0662c1b1a231-logs\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636497 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-config-data\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636516 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636535 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/549beb55-bab4-4b1a-bf98-48de2e12956f-horizon-secret-key\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636553 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636575 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-scripts\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636594 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-config-data\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636620 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/549beb55-bab4-4b1a-bf98-48de2e12956f-logs\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636637 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-db-sync-config-data\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636657 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-logs\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636671 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxqmm\" (UniqueName: \"kubernetes.io/projected/549beb55-bab4-4b1a-bf98-48de2e12956f-kube-api-access-gxqmm\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636694 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmlgz\" (UniqueName: \"kubernetes.io/projected/f1f108e1-27bb-433e-b7e1-f9cc7f778182-kube-api-access-gmlgz\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636713 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rtp8\" (UniqueName: \"kubernetes.io/projected/1227b752-8a3c-4c01-91bf-0662c1b1a231-kube-api-access-5rtp8\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636730 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqhps\" (UniqueName: \"kubernetes.io/projected/b8ad7a10-6353-45f2-a11d-126d0d4786b0-kube-api-access-kqhps\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636753 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-combined-ca-bundle\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636770 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-scripts\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636787 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636810 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-config-data\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636830 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-combined-ca-bundle\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.636856 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-scripts\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.641536 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1227b752-8a3c-4c01-91bf-0662c1b1a231-logs\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.642256 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.642574 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-logs\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.649497 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.652451 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-config-data\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.655158 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.664556 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-scripts\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.668655 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-combined-ca-bundle\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.672600 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqhps\" (UniqueName: \"kubernetes.io/projected/b8ad7a10-6353-45f2-a11d-126d0d4786b0-kube-api-access-kqhps\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.677352 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-scripts\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.678330 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-config-data\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.679344 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rtp8\" (UniqueName: \"kubernetes.io/projected/1227b752-8a3c-4c01-91bf-0662c1b1a231-kube-api-access-5rtp8\") pod \"placement-db-sync-6mz7x\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.688587 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.690927 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.695754 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.705923 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.731402 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.742671 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-config\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.742746 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.742783 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.742824 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.742847 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743147 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-config-data\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743171 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/549beb55-bab4-4b1a-bf98-48de2e12956f-horizon-secret-key\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743216 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743238 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-scripts\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743291 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743310 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/549beb55-bab4-4b1a-bf98-48de2e12956f-logs\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743327 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-db-sync-config-data\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743343 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743364 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxqmm\" (UniqueName: \"kubernetes.io/projected/549beb55-bab4-4b1a-bf98-48de2e12956f-kube-api-access-gxqmm\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743387 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svj9w\" (UniqueName: \"kubernetes.io/projected/6e945937-1e9d-465b-a68f-c755119ab6f9-kube-api-access-svj9w\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743406 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmlgz\" (UniqueName: \"kubernetes.io/projected/f1f108e1-27bb-433e-b7e1-f9cc7f778182-kube-api-access-gmlgz\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743427 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqddr\" (UniqueName: \"kubernetes.io/projected/94b82d55-b467-408f-8eb1-6da4648dde97-kube-api-access-rqddr\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743445 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-combined-ca-bundle\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743460 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743477 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.743497 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-logs\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.745796 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-config-data\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.748302 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-scripts\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.748905 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/549beb55-bab4-4b1a-bf98-48de2e12956f-logs\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.775488 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.777454 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/549beb55-bab4-4b1a-bf98-48de2e12956f-horizon-secret-key\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.778067 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-combined-ca-bundle\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.787172 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmlgz\" (UniqueName: \"kubernetes.io/projected/f1f108e1-27bb-433e-b7e1-f9cc7f778182-kube-api-access-gmlgz\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.792683 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.793882 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-db-sync-config-data\") pod \"barbican-db-sync-c7zzv\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.796466 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxqmm\" (UniqueName: \"kubernetes.io/projected/549beb55-bab4-4b1a-bf98-48de2e12956f-kube-api-access-gxqmm\") pod \"horizon-54d5f75fb9-s2659\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.844867 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.844959 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845007 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845036 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svj9w\" (UniqueName: \"kubernetes.io/projected/6e945937-1e9d-465b-a68f-c755119ab6f9-kube-api-access-svj9w\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845103 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqddr\" (UniqueName: \"kubernetes.io/projected/94b82d55-b467-408f-8eb1-6da4648dde97-kube-api-access-rqddr\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845122 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845168 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845193 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-logs\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845261 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845282 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-config\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845342 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845366 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.845407 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.846793 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.851915 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.852212 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.852636 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-logs\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.853229 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.853475 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.853747 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.856771 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-config\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.860427 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.865929 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.874919 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqddr\" (UniqueName: \"kubernetes.io/projected/94b82d55-b467-408f-8eb1-6da4648dde97-kube-api-access-rqddr\") pod \"dnsmasq-dns-785d8bcb8c-96668\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.875572 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.884948 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svj9w\" (UniqueName: \"kubernetes.io/projected/6e945937-1e9d-465b-a68f-c755119ab6f9-kube-api-access-svj9w\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.924846 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.931497 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.954043 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.968221 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:14 crc kubenswrapper[4948]: I0220 08:22:14.989095 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-g72bc"] Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.032774 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.156569 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-gmdvb"] Feb 20 08:22:15 crc kubenswrapper[4948]: W0220 08:22:15.214770 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69c38b92_a051_4834_bf34_8612edb31dfa.slice/crio-4750fdb6550fc9d53336366b666ece8cbabd6c85ec2d4df49fece59f3bc13a6a WatchSource:0}: Error finding container 4750fdb6550fc9d53336366b666ece8cbabd6c85ec2d4df49fece59f3bc13a6a: Status 404 returned error can't find the container with id 4750fdb6550fc9d53336366b666ece8cbabd6c85ec2d4df49fece59f3bc13a6a Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.218776 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d578c747f-8f6hc"] Feb 20 08:22:15 crc kubenswrapper[4948]: W0220 08:22:15.225347 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8425fd12_ae45_4e7e_af4d_9bfc3c3d0eb4.slice/crio-3f991d1b028cfdcb1512441c7603f82fee2c2bba86ea276855760a70987a1d8f WatchSource:0}: Error finding container 3f991d1b028cfdcb1512441c7603f82fee2c2bba86ea276855760a70987a1d8f: Status 404 returned error can't find the container with id 3f991d1b028cfdcb1512441c7603f82fee2c2bba86ea276855760a70987a1d8f Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.453909 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-g5zxh"] Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.464136 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-xt2zq"] Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.576024 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d578c747f-8f6hc" event={"ID":"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4","Type":"ContainerStarted","Data":"3f991d1b028cfdcb1512441c7603f82fee2c2bba86ea276855760a70987a1d8f"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.585775 4948 generic.go:334] "Generic (PLEG): container finished" podID="7960b704-21ac-4d75-8226-a9bbae4c2b50" containerID="f7f0bf783e5f62b8ff1fccd63b4e33b097f4a17ad1c512a814c4baa9c6f1eb03" exitCode=0 Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.586038 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" event={"ID":"7960b704-21ac-4d75-8226-a9bbae4c2b50","Type":"ContainerDied","Data":"f7f0bf783e5f62b8ff1fccd63b4e33b097f4a17ad1c512a814c4baa9c6f1eb03"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.588090 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.593460 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-g72bc" event={"ID":"9959c3d4-fb0e-473c-8bbe-065110a3e03a","Type":"ContainerStarted","Data":"fa275de1b67d8651c437fb815c4f8470dc8d77cd9885aaddee1a7419f33fd25c"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.593497 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-g72bc" event={"ID":"9959c3d4-fb0e-473c-8bbe-065110a3e03a","Type":"ContainerStarted","Data":"1eedce1f0a19f6576c482c3583bf052a9786f8dcd27e5bce2d34fd0ef16f70ad"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.610809 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xt2zq" event={"ID":"f706b741-86f9-49a3-95d5-85ee62eb3668","Type":"ContainerStarted","Data":"571bfb34e93031079bb93437b09bfef171c7529235d603237690b075a9933988"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.615581 4948 generic.go:334] "Generic (PLEG): container finished" podID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" containerID="cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d" exitCode=0 Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.615632 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" event={"ID":"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e","Type":"ContainerDied","Data":"cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.615656 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" event={"ID":"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e","Type":"ContainerDied","Data":"5b5cca04b9c3f166250baa067797237d67a3b5df1ae92815a7da886a955a6658"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.615674 4948 scope.go:117] "RemoveContainer" containerID="cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.615781 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-mvrcc" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.643871 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-g5zxh" event={"ID":"bb362d0e-68b4-4d48-885c-a6894ac71e6e","Type":"ContainerStarted","Data":"fb0ac2219219fe66e5a42c9ab8b34f9c1f674fc235bce862e34dedbfa83dd480"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.665706 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" event={"ID":"69c38b92-a051-4834-bf34-8612edb31dfa","Type":"ContainerStarted","Data":"545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.665744 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" event={"ID":"69c38b92-a051-4834-bf34-8612edb31dfa","Type":"ContainerStarted","Data":"4750fdb6550fc9d53336366b666ece8cbabd6c85ec2d4df49fece59f3bc13a6a"} Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.678052 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-svc\") pod \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.678109 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-config\") pod \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.678177 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57xf5\" (UniqueName: \"kubernetes.io/projected/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-kube-api-access-57xf5\") pod \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.678237 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-swift-storage-0\") pod \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.678315 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-sb\") pod \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.678344 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-nb\") pod \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\" (UID: \"0b8788ed-d6f3-4344-b8b2-7bb554ecd25e\") " Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.684386 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.691885 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-kube-api-access-57xf5" (OuterVolumeSpecName: "kube-api-access-57xf5") pod "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" (UID: "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e"). InnerVolumeSpecName "kube-api-access-57xf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.725004 4948 scope.go:117] "RemoveContainer" containerID="596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.744658 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-g72bc" podStartSLOduration=2.744634106 podStartE2EDuration="2.744634106s" podCreationTimestamp="2026-02-20 08:22:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:15.651666045 +0000 UTC m=+984.626160865" watchObservedRunningTime="2026-02-20 08:22:15.744634106 +0000 UTC m=+984.719128926" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.783130 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57xf5\" (UniqueName: \"kubernetes.io/projected/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-kube-api-access-57xf5\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.789361 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.813381 4948 scope.go:117] "RemoveContainer" containerID="cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d" Feb 20 08:22:15 crc kubenswrapper[4948]: E0220 08:22:15.816915 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d\": container with ID starting with cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d not found: ID does not exist" containerID="cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.816949 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d"} err="failed to get container status \"cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d\": rpc error: code = NotFound desc = could not find container \"cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d\": container with ID starting with cd857fd81b9d60885f23abd3b2c07fd82f462c02c36ab738b3626ed94de1c59d not found: ID does not exist" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.816986 4948 scope.go:117] "RemoveContainer" containerID="596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58" Feb 20 08:22:15 crc kubenswrapper[4948]: E0220 08:22:15.821125 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58\": container with ID starting with 596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58 not found: ID does not exist" containerID="596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.821175 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58"} err="failed to get container status \"596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58\": rpc error: code = NotFound desc = could not find container \"596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58\": container with ID starting with 596bfc5f705d539b55d6de0f0ebc270b6e2a7ef5c99c535a13f5bc566d754f58 not found: ID does not exist" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.840432 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.912823 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnp7g"] Feb 20 08:22:15 crc kubenswrapper[4948]: I0220 08:22:15.964519 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" (UID: "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.002760 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-config" (OuterVolumeSpecName: "config") pod "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" (UID: "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.002869 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.019816 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" (UID: "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.031454 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" (UID: "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.034660 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" (UID: "0b8788ed-d6f3-4344-b8b2-7bb554ecd25e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.110531 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.110567 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.110576 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.110585 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.174024 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.218501 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d578c747f-8f6hc"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.251592 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-68c9c4f5fc-2vtb4"] Feb 20 08:22:16 crc kubenswrapper[4948]: E0220 08:22:16.251964 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" containerName="dnsmasq-dns" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.251989 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" containerName="dnsmasq-dns" Feb 20 08:22:16 crc kubenswrapper[4948]: E0220 08:22:16.252009 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" containerName="init" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.252016 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" containerName="init" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.253167 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" containerName="dnsmasq-dns" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.254091 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.278925 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68c9c4f5fc-2vtb4"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.286404 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-mvrcc"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.295036 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-mvrcc"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.315451 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.321912 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-config-data\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.321988 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/136dd621-a5c5-481a-9b86-66f8cb72486d-logs\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.322015 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/136dd621-a5c5-481a-9b86-66f8cb72486d-horizon-secret-key\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.322072 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp59z\" (UniqueName: \"kubernetes.io/projected/136dd621-a5c5-481a-9b86-66f8cb72486d-kube-api-access-vp59z\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.322090 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-scripts\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.428929 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-config-data\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.429029 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/136dd621-a5c5-481a-9b86-66f8cb72486d-logs\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.429067 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/136dd621-a5c5-481a-9b86-66f8cb72486d-horizon-secret-key\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.429136 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp59z\" (UniqueName: \"kubernetes.io/projected/136dd621-a5c5-481a-9b86-66f8cb72486d-kube-api-access-vp59z\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.429159 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-scripts\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.429504 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/136dd621-a5c5-481a-9b86-66f8cb72486d-logs\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.430106 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-scripts\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.430376 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-config-data\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.466809 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/136dd621-a5c5-481a-9b86-66f8cb72486d-horizon-secret-key\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.473025 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp59z\" (UniqueName: \"kubernetes.io/projected/136dd621-a5c5-481a-9b86-66f8cb72486d-kube-api-access-vp59z\") pod \"horizon-68c9c4f5fc-2vtb4\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.514679 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.527863 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.556885 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.570928 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-c7zzv"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.591871 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-54d5f75fb9-s2659"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.604211 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.621530 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-96668"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.636741 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-svc\") pod \"7960b704-21ac-4d75-8226-a9bbae4c2b50\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.636921 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-svc\") pod \"69c38b92-a051-4834-bf34-8612edb31dfa\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.636958 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28ln7\" (UniqueName: \"kubernetes.io/projected/69c38b92-a051-4834-bf34-8612edb31dfa-kube-api-access-28ln7\") pod \"69c38b92-a051-4834-bf34-8612edb31dfa\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639315 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-sb\") pod \"7960b704-21ac-4d75-8226-a9bbae4c2b50\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639353 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-nb\") pod \"7960b704-21ac-4d75-8226-a9bbae4c2b50\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639397 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-config\") pod \"7960b704-21ac-4d75-8226-a9bbae4c2b50\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639440 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-nb\") pod \"69c38b92-a051-4834-bf34-8612edb31dfa\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639469 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-swift-storage-0\") pod \"69c38b92-a051-4834-bf34-8612edb31dfa\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639491 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-sb\") pod \"69c38b92-a051-4834-bf34-8612edb31dfa\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639534 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74rwq\" (UniqueName: \"kubernetes.io/projected/7960b704-21ac-4d75-8226-a9bbae4c2b50-kube-api-access-74rwq\") pod \"7960b704-21ac-4d75-8226-a9bbae4c2b50\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639552 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-swift-storage-0\") pod \"7960b704-21ac-4d75-8226-a9bbae4c2b50\" (UID: \"7960b704-21ac-4d75-8226-a9bbae4c2b50\") " Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.639595 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-config\") pod \"69c38b92-a051-4834-bf34-8612edb31dfa\" (UID: \"69c38b92-a051-4834-bf34-8612edb31dfa\") " Feb 20 08:22:16 crc kubenswrapper[4948]: W0220 08:22:16.645724 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1227b752_8a3c_4c01_91bf_0662c1b1a231.slice/crio-5b0f5c257994604a44a9cf0e2c06abc610dcd1eb5c3d34cfd415bfd74671d08f WatchSource:0}: Error finding container 5b0f5c257994604a44a9cf0e2c06abc610dcd1eb5c3d34cfd415bfd74671d08f: Status 404 returned error can't find the container with id 5b0f5c257994604a44a9cf0e2c06abc610dcd1eb5c3d34cfd415bfd74671d08f Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.652593 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69c38b92-a051-4834-bf34-8612edb31dfa-kube-api-access-28ln7" (OuterVolumeSpecName: "kube-api-access-28ln7") pod "69c38b92-a051-4834-bf34-8612edb31dfa" (UID: "69c38b92-a051-4834-bf34-8612edb31dfa"). InnerVolumeSpecName "kube-api-access-28ln7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.673911 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54d5f75fb9-s2659" event={"ID":"549beb55-bab4-4b1a-bf98-48de2e12956f","Type":"ContainerStarted","Data":"27986056f94b762ad4c0ec91cfd1562122b907c3b47800656945a8588571afe8"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.677013 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerStarted","Data":"2c29cb12a24e0769e2fc11186d9dff0648d1bcb261fcbab7060faff8fb76085a"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.684808 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6mz7x"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.690851 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7960b704-21ac-4d75-8226-a9bbae4c2b50-kube-api-access-74rwq" (OuterVolumeSpecName: "kube-api-access-74rwq") pod "7960b704-21ac-4d75-8226-a9bbae4c2b50" (UID: "7960b704-21ac-4d75-8226-a9bbae4c2b50"). InnerVolumeSpecName "kube-api-access-74rwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.702360 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" event={"ID":"94b82d55-b467-408f-8eb1-6da4648dde97","Type":"ContainerStarted","Data":"af3aca3908294e0c1fe45deb6418a29e5d516dc4d54d54ff11acdb6845abd301"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.705002 4948 generic.go:334] "Generic (PLEG): container finished" podID="69c38b92-a051-4834-bf34-8612edb31dfa" containerID="545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4" exitCode=0 Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.705038 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" event={"ID":"69c38b92-a051-4834-bf34-8612edb31dfa","Type":"ContainerDied","Data":"545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.705054 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" event={"ID":"69c38b92-a051-4834-bf34-8612edb31dfa","Type":"ContainerDied","Data":"4750fdb6550fc9d53336366b666ece8cbabd6c85ec2d4df49fece59f3bc13a6a"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.705069 4948 scope.go:117] "RemoveContainer" containerID="545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.705177 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-gmdvb" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.706839 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" event={"ID":"7960b704-21ac-4d75-8226-a9bbae4c2b50","Type":"ContainerDied","Data":"e3cb32ae5e929563597aaafd43d3401d85066f36163aa0afc9720c6e975f05be"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.706903 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-v6srk" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.713245 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.714195 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8ad7a10-6353-45f2-a11d-126d0d4786b0","Type":"ContainerStarted","Data":"434c4fc6de7f62028433ff88fb8c531a01a506655fd56b116472009d2d35dd9f"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.716405 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-g5zxh" event={"ID":"bb362d0e-68b4-4d48-885c-a6894ac71e6e","Type":"ContainerStarted","Data":"cb7241154041475439806eb631212532adb6e6775adf70e62bdf5cf4b6d19276"} Feb 20 08:22:16 crc kubenswrapper[4948]: W0220 08:22:16.724418 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e945937_1e9d_465b_a68f_c755119ab6f9.slice/crio-feaec65599576490d6ae435b0c4ffac29904876fc82b7f96fd44c9df4357c090 WatchSource:0}: Error finding container feaec65599576490d6ae435b0c4ffac29904876fc82b7f96fd44c9df4357c090: Status 404 returned error can't find the container with id feaec65599576490d6ae435b0c4ffac29904876fc82b7f96fd44c9df4357c090 Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.725574 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6mz7x" event={"ID":"1227b752-8a3c-4c01-91bf-0662c1b1a231","Type":"ContainerStarted","Data":"5b0f5c257994604a44a9cf0e2c06abc610dcd1eb5c3d34cfd415bfd74671d08f"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.731754 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-c7zzv" event={"ID":"f1f108e1-27bb-433e-b7e1-f9cc7f778182","Type":"ContainerStarted","Data":"1c7c890826add7e27d4d21f7cb357370f931db1e77090c60ac1c40f2ed6d57cc"} Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.735385 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-g5zxh" podStartSLOduration=3.735356072 podStartE2EDuration="3.735356072s" podCreationTimestamp="2026-02-20 08:22:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:16.731488736 +0000 UTC m=+985.705983556" watchObservedRunningTime="2026-02-20 08:22:16.735356072 +0000 UTC m=+985.709850892" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.745163 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28ln7\" (UniqueName: \"kubernetes.io/projected/69c38b92-a051-4834-bf34-8612edb31dfa-kube-api-access-28ln7\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.745195 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74rwq\" (UniqueName: \"kubernetes.io/projected/7960b704-21ac-4d75-8226-a9bbae4c2b50-kube-api-access-74rwq\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.754566 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-config" (OuterVolumeSpecName: "config") pod "69c38b92-a051-4834-bf34-8612edb31dfa" (UID: "69c38b92-a051-4834-bf34-8612edb31dfa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.764238 4948 scope.go:117] "RemoveContainer" containerID="545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4" Feb 20 08:22:16 crc kubenswrapper[4948]: E0220 08:22:16.766172 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4\": container with ID starting with 545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4 not found: ID does not exist" containerID="545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.766203 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4"} err="failed to get container status \"545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4\": rpc error: code = NotFound desc = could not find container \"545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4\": container with ID starting with 545ede59ed041c4b028c3e626eff3b4e85ad826a2ef8b2830cd52655853e1ad4 not found: ID does not exist" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.766221 4948 scope.go:117] "RemoveContainer" containerID="f7f0bf783e5f62b8ff1fccd63b4e33b097f4a17ad1c512a814c4baa9c6f1eb03" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.856997 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.894336 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "69c38b92-a051-4834-bf34-8612edb31dfa" (UID: "69c38b92-a051-4834-bf34-8612edb31dfa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.900352 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7960b704-21ac-4d75-8226-a9bbae4c2b50" (UID: "7960b704-21ac-4d75-8226-a9bbae4c2b50"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.909659 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7960b704-21ac-4d75-8226-a9bbae4c2b50" (UID: "7960b704-21ac-4d75-8226-a9bbae4c2b50"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.938043 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7960b704-21ac-4d75-8226-a9bbae4c2b50" (UID: "7960b704-21ac-4d75-8226-a9bbae4c2b50"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.941248 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "69c38b92-a051-4834-bf34-8612edb31dfa" (UID: "69c38b92-a051-4834-bf34-8612edb31dfa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.943884 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "69c38b92-a051-4834-bf34-8612edb31dfa" (UID: "69c38b92-a051-4834-bf34-8612edb31dfa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.951930 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "69c38b92-a051-4834-bf34-8612edb31dfa" (UID: "69c38b92-a051-4834-bf34-8612edb31dfa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.958326 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.958346 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.958354 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.958364 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.958372 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.958380 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69c38b92-a051-4834-bf34-8612edb31dfa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.958388 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.960287 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-config" (OuterVolumeSpecName: "config") pod "7960b704-21ac-4d75-8226-a9bbae4c2b50" (UID: "7960b704-21ac-4d75-8226-a9bbae4c2b50"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:16 crc kubenswrapper[4948]: I0220 08:22:16.973483 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7960b704-21ac-4d75-8226-a9bbae4c2b50" (UID: "7960b704-21ac-4d75-8226-a9bbae4c2b50"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.059922 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.059954 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7960b704-21ac-4d75-8226-a9bbae4c2b50-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.195190 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-gmdvb"] Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.243949 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-gmdvb"] Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.266231 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-v6srk"] Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.275082 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-v6srk"] Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.281609 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68c9c4f5fc-2vtb4"] Feb 20 08:22:17 crc kubenswrapper[4948]: E0220 08:22:17.513362 4948 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94b82d55_b467_408f_8eb1_6da4648dde97.slice/crio-bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94b82d55_b467_408f_8eb1_6da4648dde97.slice/crio-conmon-bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df.scope\": RecentStats: unable to find data in memory cache]" Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.743449 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b8788ed-d6f3-4344-b8b2-7bb554ecd25e" path="/var/lib/kubelet/pods/0b8788ed-d6f3-4344-b8b2-7bb554ecd25e/volumes" Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.744414 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69c38b92-a051-4834-bf34-8612edb31dfa" path="/var/lib/kubelet/pods/69c38b92-a051-4834-bf34-8612edb31dfa/volumes" Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.745381 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7960b704-21ac-4d75-8226-a9bbae4c2b50" path="/var/lib/kubelet/pods/7960b704-21ac-4d75-8226-a9bbae4c2b50/volumes" Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.789094 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8ad7a10-6353-45f2-a11d-126d0d4786b0","Type":"ContainerStarted","Data":"fdfa9ea96d17b0fbac466add147012945b8f9c77bfe4c7e3939115948c20ff15"} Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.797186 4948 generic.go:334] "Generic (PLEG): container finished" podID="94b82d55-b467-408f-8eb1-6da4648dde97" containerID="bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df" exitCode=0 Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.797270 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" event={"ID":"94b82d55-b467-408f-8eb1-6da4648dde97","Type":"ContainerDied","Data":"bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df"} Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.804742 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e945937-1e9d-465b-a68f-c755119ab6f9","Type":"ContainerStarted","Data":"feaec65599576490d6ae435b0c4ffac29904876fc82b7f96fd44c9df4357c090"} Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.817451 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c9c4f5fc-2vtb4" event={"ID":"136dd621-a5c5-481a-9b86-66f8cb72486d","Type":"ContainerStarted","Data":"45d6073bbc41d12c6fd4a83a0736dfc692eb915e2f643d37fcfcdd4574770ee8"} Feb 20 08:22:17 crc kubenswrapper[4948]: I0220 08:22:17.819008 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lnp7g" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerName="registry-server" containerID="cri-o://59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68" gracePeriod=2 Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.495142 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.615470 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-catalog-content\") pod \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.615529 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-utilities\") pod \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.615551 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5vn9\" (UniqueName: \"kubernetes.io/projected/81e7d5de-2379-46c7-ba9b-ffcacad79e29-kube-api-access-w5vn9\") pod \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\" (UID: \"81e7d5de-2379-46c7-ba9b-ffcacad79e29\") " Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.617143 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-utilities" (OuterVolumeSpecName: "utilities") pod "81e7d5de-2379-46c7-ba9b-ffcacad79e29" (UID: "81e7d5de-2379-46c7-ba9b-ffcacad79e29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.625716 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81e7d5de-2379-46c7-ba9b-ffcacad79e29-kube-api-access-w5vn9" (OuterVolumeSpecName: "kube-api-access-w5vn9") pod "81e7d5de-2379-46c7-ba9b-ffcacad79e29" (UID: "81e7d5de-2379-46c7-ba9b-ffcacad79e29"). InnerVolumeSpecName "kube-api-access-w5vn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.650957 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "81e7d5de-2379-46c7-ba9b-ffcacad79e29" (UID: "81e7d5de-2379-46c7-ba9b-ffcacad79e29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.717973 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.718238 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e7d5de-2379-46c7-ba9b-ffcacad79e29-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.718255 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5vn9\" (UniqueName: \"kubernetes.io/projected/81e7d5de-2379-46c7-ba9b-ffcacad79e29-kube-api-access-w5vn9\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.837283 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnp7g" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.837584 4948 generic.go:334] "Generic (PLEG): container finished" podID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerID="59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68" exitCode=0 Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.837666 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnp7g" event={"ID":"81e7d5de-2379-46c7-ba9b-ffcacad79e29","Type":"ContainerDied","Data":"59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68"} Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.837692 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnp7g" event={"ID":"81e7d5de-2379-46c7-ba9b-ffcacad79e29","Type":"ContainerDied","Data":"ac4ceb2acdc56fbdc4b52d619a840394196af858618bbcf55bc7c1dc060d1a2f"} Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.837708 4948 scope.go:117] "RemoveContainer" containerID="59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.842544 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8ad7a10-6353-45f2-a11d-126d0d4786b0","Type":"ContainerStarted","Data":"b74656e29439723075b3c2514dc8006478ea22c951aec141fc385a6ef5cf24b1"} Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.842661 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerName="glance-log" containerID="cri-o://fdfa9ea96d17b0fbac466add147012945b8f9c77bfe4c7e3939115948c20ff15" gracePeriod=30 Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.842809 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerName="glance-httpd" containerID="cri-o://b74656e29439723075b3c2514dc8006478ea22c951aec141fc385a6ef5cf24b1" gracePeriod=30 Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.855417 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" event={"ID":"94b82d55-b467-408f-8eb1-6da4648dde97","Type":"ContainerStarted","Data":"76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0"} Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.855573 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.858773 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e945937-1e9d-465b-a68f-c755119ab6f9","Type":"ContainerStarted","Data":"e38f809977ba5076c8af037b7e38d52d33545f893ce90480e346caeaaef81dee"} Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.906507 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.90648207 podStartE2EDuration="4.90648207s" podCreationTimestamp="2026-02-20 08:22:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:18.900492322 +0000 UTC m=+987.874987142" watchObservedRunningTime="2026-02-20 08:22:18.90648207 +0000 UTC m=+987.880976890" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.916689 4948 scope.go:117] "RemoveContainer" containerID="833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.943311 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" podStartSLOduration=4.943287031 podStartE2EDuration="4.943287031s" podCreationTimestamp="2026-02-20 08:22:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:18.92670419 +0000 UTC m=+987.901199030" watchObservedRunningTime="2026-02-20 08:22:18.943287031 +0000 UTC m=+987.917781851" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.973856 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnp7g"] Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.975517 4948 scope.go:117] "RemoveContainer" containerID="82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b" Feb 20 08:22:18 crc kubenswrapper[4948]: I0220 08:22:18.979341 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnp7g"] Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.073130 4948 scope.go:117] "RemoveContainer" containerID="59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68" Feb 20 08:22:19 crc kubenswrapper[4948]: E0220 08:22:19.075728 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68\": container with ID starting with 59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68 not found: ID does not exist" containerID="59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68" Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.075769 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68"} err="failed to get container status \"59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68\": rpc error: code = NotFound desc = could not find container \"59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68\": container with ID starting with 59eceb81df286be225188d13264d44596719c0a06ca1c39903c517b422013a68 not found: ID does not exist" Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.075792 4948 scope.go:117] "RemoveContainer" containerID="833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15" Feb 20 08:22:19 crc kubenswrapper[4948]: E0220 08:22:19.076360 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15\": container with ID starting with 833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15 not found: ID does not exist" containerID="833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15" Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.076381 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15"} err="failed to get container status \"833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15\": rpc error: code = NotFound desc = could not find container \"833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15\": container with ID starting with 833390a5ef5c4525790a7cad88740d3ce4d0df639a0829042164aca3b32d8d15 not found: ID does not exist" Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.076401 4948 scope.go:117] "RemoveContainer" containerID="82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b" Feb 20 08:22:19 crc kubenswrapper[4948]: E0220 08:22:19.076886 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b\": container with ID starting with 82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b not found: ID does not exist" containerID="82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b" Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.076905 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b"} err="failed to get container status \"82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b\": rpc error: code = NotFound desc = could not find container \"82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b\": container with ID starting with 82db4fd040d5ae64aa60b6299d284256aa892181e4cf858ebfad62dc9681817b not found: ID does not exist" Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.742076 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" path="/var/lib/kubelet/pods/81e7d5de-2379-46c7-ba9b-ffcacad79e29/volumes" Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.878445 4948 generic.go:334] "Generic (PLEG): container finished" podID="9959c3d4-fb0e-473c-8bbe-065110a3e03a" containerID="fa275de1b67d8651c437fb815c4f8470dc8d77cd9885aaddee1a7419f33fd25c" exitCode=0 Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.878514 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-g72bc" event={"ID":"9959c3d4-fb0e-473c-8bbe-065110a3e03a","Type":"ContainerDied","Data":"fa275de1b67d8651c437fb815c4f8470dc8d77cd9885aaddee1a7419f33fd25c"} Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.885409 4948 generic.go:334] "Generic (PLEG): container finished" podID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerID="b74656e29439723075b3c2514dc8006478ea22c951aec141fc385a6ef5cf24b1" exitCode=0 Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.885442 4948 generic.go:334] "Generic (PLEG): container finished" podID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerID="fdfa9ea96d17b0fbac466add147012945b8f9c77bfe4c7e3939115948c20ff15" exitCode=143 Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.885513 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8ad7a10-6353-45f2-a11d-126d0d4786b0","Type":"ContainerDied","Data":"b74656e29439723075b3c2514dc8006478ea22c951aec141fc385a6ef5cf24b1"} Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.885542 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8ad7a10-6353-45f2-a11d-126d0d4786b0","Type":"ContainerDied","Data":"fdfa9ea96d17b0fbac466add147012945b8f9c77bfe4c7e3939115948c20ff15"} Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.889623 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e945937-1e9d-465b-a68f-c755119ab6f9","Type":"ContainerStarted","Data":"344f41da66c91718be93d872d218e57c352e04b55c97d3dc69d71b942dbe58d1"} Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.890037 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerName="glance-log" containerID="cri-o://e38f809977ba5076c8af037b7e38d52d33545f893ce90480e346caeaaef81dee" gracePeriod=30 Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.890185 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerName="glance-httpd" containerID="cri-o://344f41da66c91718be93d872d218e57c352e04b55c97d3dc69d71b942dbe58d1" gracePeriod=30 Feb 20 08:22:19 crc kubenswrapper[4948]: I0220 08:22:19.926328 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.926307868 podStartE2EDuration="5.926307868s" podCreationTimestamp="2026-02-20 08:22:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:19.920483574 +0000 UTC m=+988.894978394" watchObservedRunningTime="2026-02-20 08:22:19.926307868 +0000 UTC m=+988.900802688" Feb 20 08:22:20 crc kubenswrapper[4948]: I0220 08:22:20.904558 4948 generic.go:334] "Generic (PLEG): container finished" podID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerID="344f41da66c91718be93d872d218e57c352e04b55c97d3dc69d71b942dbe58d1" exitCode=0 Feb 20 08:22:20 crc kubenswrapper[4948]: I0220 08:22:20.905557 4948 generic.go:334] "Generic (PLEG): container finished" podID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerID="e38f809977ba5076c8af037b7e38d52d33545f893ce90480e346caeaaef81dee" exitCode=143 Feb 20 08:22:20 crc kubenswrapper[4948]: I0220 08:22:20.904904 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e945937-1e9d-465b-a68f-c755119ab6f9","Type":"ContainerDied","Data":"344f41da66c91718be93d872d218e57c352e04b55c97d3dc69d71b942dbe58d1"} Feb 20 08:22:20 crc kubenswrapper[4948]: I0220 08:22:20.905814 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e945937-1e9d-465b-a68f-c755119ab6f9","Type":"ContainerDied","Data":"e38f809977ba5076c8af037b7e38d52d33545f893ce90480e346caeaaef81dee"} Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.347668 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.452475 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-scripts\") pod \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.452564 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqhps\" (UniqueName: \"kubernetes.io/projected/b8ad7a10-6353-45f2-a11d-126d0d4786b0-kube-api-access-kqhps\") pod \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.452606 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-logs\") pod \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.452678 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-httpd-run\") pod \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.452842 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-combined-ca-bundle\") pod \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.452893 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.452950 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-config-data\") pod \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\" (UID: \"b8ad7a10-6353-45f2-a11d-126d0d4786b0\") " Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.453043 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-logs" (OuterVolumeSpecName: "logs") pod "b8ad7a10-6353-45f2-a11d-126d0d4786b0" (UID: "b8ad7a10-6353-45f2-a11d-126d0d4786b0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.453115 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b8ad7a10-6353-45f2-a11d-126d0d4786b0" (UID: "b8ad7a10-6353-45f2-a11d-126d0d4786b0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.453398 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.453420 4948 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8ad7a10-6353-45f2-a11d-126d0d4786b0-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.459662 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8ad7a10-6353-45f2-a11d-126d0d4786b0-kube-api-access-kqhps" (OuterVolumeSpecName: "kube-api-access-kqhps") pod "b8ad7a10-6353-45f2-a11d-126d0d4786b0" (UID: "b8ad7a10-6353-45f2-a11d-126d0d4786b0"). InnerVolumeSpecName "kube-api-access-kqhps". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.460125 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-scripts" (OuterVolumeSpecName: "scripts") pod "b8ad7a10-6353-45f2-a11d-126d0d4786b0" (UID: "b8ad7a10-6353-45f2-a11d-126d0d4786b0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.460445 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "b8ad7a10-6353-45f2-a11d-126d0d4786b0" (UID: "b8ad7a10-6353-45f2-a11d-126d0d4786b0"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.499564 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8ad7a10-6353-45f2-a11d-126d0d4786b0" (UID: "b8ad7a10-6353-45f2-a11d-126d0d4786b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.520145 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-config-data" (OuterVolumeSpecName: "config-data") pod "b8ad7a10-6353-45f2-a11d-126d0d4786b0" (UID: "b8ad7a10-6353-45f2-a11d-126d0d4786b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.555201 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqhps\" (UniqueName: \"kubernetes.io/projected/b8ad7a10-6353-45f2-a11d-126d0d4786b0-kube-api-access-kqhps\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.555246 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.555288 4948 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.555304 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.555317 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8ad7a10-6353-45f2-a11d-126d0d4786b0-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.577402 4948 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.658405 4948 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.937212 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8ad7a10-6353-45f2-a11d-126d0d4786b0","Type":"ContainerDied","Data":"434c4fc6de7f62028433ff88fb8c531a01a506655fd56b116472009d2d35dd9f"} Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.937296 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.937382 4948 scope.go:117] "RemoveContainer" containerID="b74656e29439723075b3c2514dc8006478ea22c951aec141fc385a6ef5cf24b1" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.970516 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.978284 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.997840 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:22 crc kubenswrapper[4948]: E0220 08:22:22.998361 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69c38b92-a051-4834-bf34-8612edb31dfa" containerName="init" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998380 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="69c38b92-a051-4834-bf34-8612edb31dfa" containerName="init" Feb 20 08:22:22 crc kubenswrapper[4948]: E0220 08:22:22.998415 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerName="glance-log" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998429 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerName="glance-log" Feb 20 08:22:22 crc kubenswrapper[4948]: E0220 08:22:22.998444 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7960b704-21ac-4d75-8226-a9bbae4c2b50" containerName="init" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998450 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="7960b704-21ac-4d75-8226-a9bbae4c2b50" containerName="init" Feb 20 08:22:22 crc kubenswrapper[4948]: E0220 08:22:22.998458 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerName="registry-server" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998464 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerName="registry-server" Feb 20 08:22:22 crc kubenswrapper[4948]: E0220 08:22:22.998511 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerName="glance-httpd" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998519 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerName="glance-httpd" Feb 20 08:22:22 crc kubenswrapper[4948]: E0220 08:22:22.998530 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerName="extract-utilities" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998537 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerName="extract-utilities" Feb 20 08:22:22 crc kubenswrapper[4948]: E0220 08:22:22.998553 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerName="extract-content" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998560 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerName="extract-content" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998770 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerName="glance-log" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998789 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="81e7d5de-2379-46c7-ba9b-ffcacad79e29" containerName="registry-server" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998801 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="69c38b92-a051-4834-bf34-8612edb31dfa" containerName="init" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998836 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="7960b704-21ac-4d75-8226-a9bbae4c2b50" containerName="init" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.998845 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" containerName="glance-httpd" Feb 20 08:22:22 crc kubenswrapper[4948]: I0220 08:22:22.999961 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.002146 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.016517 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.066617 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5zcb\" (UniqueName: \"kubernetes.io/projected/e0f82f51-8a20-402d-8861-6c57a4e5af0d-kube-api-access-q5zcb\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.066875 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-logs\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.066996 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.067094 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-config-data\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.067205 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-scripts\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.067318 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.067420 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.169005 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5zcb\" (UniqueName: \"kubernetes.io/projected/e0f82f51-8a20-402d-8861-6c57a4e5af0d-kube-api-access-q5zcb\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.169399 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-logs\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.169434 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.169458 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-config-data\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.169492 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-scripts\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.169524 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.169560 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.170186 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.170808 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.171698 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-logs\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.174712 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-scripts\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.176528 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.176868 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-config-data\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.196602 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5zcb\" (UniqueName: \"kubernetes.io/projected/e0f82f51-8a20-402d-8861-6c57a4e5af0d-kube-api-access-q5zcb\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.209697 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.328626 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:23 crc kubenswrapper[4948]: I0220 08:22:23.735647 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8ad7a10-6353-45f2-a11d-126d0d4786b0" path="/var/lib/kubelet/pods/b8ad7a10-6353-45f2-a11d-126d0d4786b0/volumes" Feb 20 08:22:24 crc kubenswrapper[4948]: I0220 08:22:24.712376 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:24 crc kubenswrapper[4948]: I0220 08:22:24.970196 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.045053 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tkqzb"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.046812 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-tkqzb" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="dnsmasq-dns" containerID="cri-o://562bf68215847fe83f99073d009b1ee89904610f3d702b2ab282baecbca78fb2" gracePeriod=10 Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.074775 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-54d5f75fb9-s2659"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.125462 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-66b4d4cb8-hgmkd"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.129906 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.138018 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.156883 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66b4d4cb8-hgmkd"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.228639 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68c9c4f5fc-2vtb4"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.300285 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-685c4db87b-rw57h"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.301784 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.305913 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-scripts\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.306009 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-secret-key\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.306040 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-config-data\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.306060 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-tls-certs\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.306084 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgwfv\" (UniqueName: \"kubernetes.io/projected/d1046503-c6d3-49bf-9523-2f897fcb82ba-kube-api-access-mgwfv\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.306125 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1046503-c6d3-49bf-9523-2f897fcb82ba-logs\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.306175 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-combined-ca-bundle\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.320719 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-685c4db87b-rw57h"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.407891 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b9a072a-a040-414c-96c7-cf1454148745-scripts\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408204 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7b9a072a-a040-414c-96c7-cf1454148745-config-data\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408299 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-scripts\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408374 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-combined-ca-bundle\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408536 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-secret-key\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408585 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-config-data\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408605 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-tls-certs\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408629 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgwfv\" (UniqueName: \"kubernetes.io/projected/d1046503-c6d3-49bf-9523-2f897fcb82ba-kube-api-access-mgwfv\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408672 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdmtj\" (UniqueName: \"kubernetes.io/projected/7b9a072a-a040-414c-96c7-cf1454148745-kube-api-access-vdmtj\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408706 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-horizon-tls-certs\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408729 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1046503-c6d3-49bf-9523-2f897fcb82ba-logs\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408830 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b9a072a-a040-414c-96c7-cf1454148745-logs\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408867 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-combined-ca-bundle\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.408930 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-horizon-secret-key\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.409466 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-scripts\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.413831 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-config-data\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.419843 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1046503-c6d3-49bf-9523-2f897fcb82ba-logs\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.421481 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-secret-key\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.421941 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-combined-ca-bundle\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.422610 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-tls-certs\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.449528 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgwfv\" (UniqueName: \"kubernetes.io/projected/d1046503-c6d3-49bf-9523-2f897fcb82ba-kube-api-access-mgwfv\") pod \"horizon-66b4d4cb8-hgmkd\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.476606 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.510445 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7b9a072a-a040-414c-96c7-cf1454148745-config-data\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.510495 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-combined-ca-bundle\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.510540 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdmtj\" (UniqueName: \"kubernetes.io/projected/7b9a072a-a040-414c-96c7-cf1454148745-kube-api-access-vdmtj\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.510562 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-horizon-tls-certs\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.510591 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b9a072a-a040-414c-96c7-cf1454148745-logs\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.510626 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-horizon-secret-key\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.510676 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b9a072a-a040-414c-96c7-cf1454148745-scripts\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.511363 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b9a072a-a040-414c-96c7-cf1454148745-scripts\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.512201 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7b9a072a-a040-414c-96c7-cf1454148745-config-data\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.514152 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b9a072a-a040-414c-96c7-cf1454148745-logs\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.518712 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-horizon-tls-certs\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.519441 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-combined-ca-bundle\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.534454 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7b9a072a-a040-414c-96c7-cf1454148745-horizon-secret-key\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.542712 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdmtj\" (UniqueName: \"kubernetes.io/projected/7b9a072a-a040-414c-96c7-cf1454148745-kube-api-access-vdmtj\") pod \"horizon-685c4db87b-rw57h\" (UID: \"7b9a072a-a040-414c-96c7-cf1454148745\") " pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.626222 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g27hk"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.627789 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.637281 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g27hk"] Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.642339 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.815515 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-catalog-content\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.815599 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2k5c\" (UniqueName: \"kubernetes.io/projected/d6729921-0a12-46a0-824f-810df88ae426-kube-api-access-x2k5c\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.815991 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-utilities\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.918169 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-utilities\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.918238 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-catalog-content\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.918302 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2k5c\" (UniqueName: \"kubernetes.io/projected/d6729921-0a12-46a0-824f-810df88ae426-kube-api-access-x2k5c\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.918759 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-utilities\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.918815 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-catalog-content\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.937550 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2k5c\" (UniqueName: \"kubernetes.io/projected/d6729921-0a12-46a0-824f-810df88ae426-kube-api-access-x2k5c\") pod \"community-operators-g27hk\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.965398 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.974359 4948 generic.go:334] "Generic (PLEG): container finished" podID="da816a64-160f-4a5c-b945-b8fa467208d6" containerID="562bf68215847fe83f99073d009b1ee89904610f3d702b2ab282baecbca78fb2" exitCode=0 Feb 20 08:22:25 crc kubenswrapper[4948]: I0220 08:22:25.974404 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tkqzb" event={"ID":"da816a64-160f-4a5c-b945-b8fa467208d6","Type":"ContainerDied","Data":"562bf68215847fe83f99073d009b1ee89904610f3d702b2ab282baecbca78fb2"} Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.349965 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.543773 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-combined-ca-bundle\") pod \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.543856 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p27c6\" (UniqueName: \"kubernetes.io/projected/9959c3d4-fb0e-473c-8bbe-065110a3e03a-kube-api-access-p27c6\") pod \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.543926 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-credential-keys\") pod \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.543964 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-scripts\") pod \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.544009 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-config-data\") pod \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.544033 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-fernet-keys\") pod \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\" (UID: \"9959c3d4-fb0e-473c-8bbe-065110a3e03a\") " Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.549731 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9959c3d4-fb0e-473c-8bbe-065110a3e03a" (UID: "9959c3d4-fb0e-473c-8bbe-065110a3e03a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.549759 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-scripts" (OuterVolumeSpecName: "scripts") pod "9959c3d4-fb0e-473c-8bbe-065110a3e03a" (UID: "9959c3d4-fb0e-473c-8bbe-065110a3e03a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.550273 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9959c3d4-fb0e-473c-8bbe-065110a3e03a-kube-api-access-p27c6" (OuterVolumeSpecName: "kube-api-access-p27c6") pod "9959c3d4-fb0e-473c-8bbe-065110a3e03a" (UID: "9959c3d4-fb0e-473c-8bbe-065110a3e03a"). InnerVolumeSpecName "kube-api-access-p27c6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.550554 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9959c3d4-fb0e-473c-8bbe-065110a3e03a" (UID: "9959c3d4-fb0e-473c-8bbe-065110a3e03a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.569079 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9959c3d4-fb0e-473c-8bbe-065110a3e03a" (UID: "9959c3d4-fb0e-473c-8bbe-065110a3e03a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.588870 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-config-data" (OuterVolumeSpecName: "config-data") pod "9959c3d4-fb0e-473c-8bbe-065110a3e03a" (UID: "9959c3d4-fb0e-473c-8bbe-065110a3e03a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.645685 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.646044 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p27c6\" (UniqueName: \"kubernetes.io/projected/9959c3d4-fb0e-473c-8bbe-065110a3e03a-kube-api-access-p27c6\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.646062 4948 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.646074 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.646086 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.646096 4948 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9959c3d4-fb0e-473c-8bbe-065110a3e03a-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.994828 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-g72bc" event={"ID":"9959c3d4-fb0e-473c-8bbe-065110a3e03a","Type":"ContainerDied","Data":"1eedce1f0a19f6576c482c3583bf052a9786f8dcd27e5bce2d34fd0ef16f70ad"} Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.994876 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1eedce1f0a19f6576c482c3583bf052a9786f8dcd27e5bce2d34fd0ef16f70ad" Feb 20 08:22:27 crc kubenswrapper[4948]: I0220 08:22:27.994954 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-g72bc" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.434314 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-g72bc"] Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.442021 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-g72bc"] Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.535304 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-v29fx"] Feb 20 08:22:28 crc kubenswrapper[4948]: E0220 08:22:28.535935 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9959c3d4-fb0e-473c-8bbe-065110a3e03a" containerName="keystone-bootstrap" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.536062 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="9959c3d4-fb0e-473c-8bbe-065110a3e03a" containerName="keystone-bootstrap" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.536387 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="9959c3d4-fb0e-473c-8bbe-065110a3e03a" containerName="keystone-bootstrap" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.537310 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.540273 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.540551 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-v9rj2" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.540684 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.541101 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.541623 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.545387 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v29fx"] Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.581822 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-scripts\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.581871 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-credential-keys\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.581907 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-config-data\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.581949 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-fernet-keys\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.582068 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-combined-ca-bundle\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.582117 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jr9d\" (UniqueName: \"kubernetes.io/projected/d84a71b6-7503-4150-a5c4-4579b08a669a-kube-api-access-6jr9d\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.684153 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-config-data\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.684243 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-fernet-keys\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.684280 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-combined-ca-bundle\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.684339 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jr9d\" (UniqueName: \"kubernetes.io/projected/d84a71b6-7503-4150-a5c4-4579b08a669a-kube-api-access-6jr9d\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.684428 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-scripts\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.684469 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-credential-keys\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.691760 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-config-data\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.696453 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-fernet-keys\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.701517 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-scripts\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.702041 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-combined-ca-bundle\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.703728 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-credential-keys\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.706380 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jr9d\" (UniqueName: \"kubernetes.io/projected/d84a71b6-7503-4150-a5c4-4579b08a669a-kube-api-access-6jr9d\") pod \"keystone-bootstrap-v29fx\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:28 crc kubenswrapper[4948]: I0220 08:22:28.866921 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:28.999704 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hh5s2"] Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.002917 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.015136 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hh5s2"] Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.095213 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-utilities\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.095548 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw6bb\" (UniqueName: \"kubernetes.io/projected/426b37f9-4acb-4d28-a55c-3fcde2600002-kube-api-access-tw6bb\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.095702 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-catalog-content\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.199741 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw6bb\" (UniqueName: \"kubernetes.io/projected/426b37f9-4acb-4d28-a55c-3fcde2600002-kube-api-access-tw6bb\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.199809 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-catalog-content\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.199873 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-utilities\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.200369 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-utilities\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.200893 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-catalog-content\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.220886 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw6bb\" (UniqueName: \"kubernetes.io/projected/426b37f9-4acb-4d28-a55c-3fcde2600002-kube-api-access-tw6bb\") pod \"certified-operators-hh5s2\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.358061 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.732132 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9959c3d4-fb0e-473c-8bbe-065110a3e03a" path="/var/lib/kubelet/pods/9959c3d4-fb0e-473c-8bbe-065110a3e03a/volumes" Feb 20 08:22:29 crc kubenswrapper[4948]: I0220 08:22:29.848630 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-tkqzb" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.527602 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.529307 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nb6h656h698h57bh5dbhcbh5bh567h57dh56ch99hb4h68fh58h597h9fh5d5h55fh66ch554h57dhbdh67bh8dh56ch567h8chb8h58fh74h5c6h656q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vp59z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-68c9c4f5fc-2vtb4_openstack(136dd621-a5c5-481a-9b86-66f8cb72486d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.532557 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-68c9c4f5fc-2vtb4" podUID="136dd621-a5c5-481a-9b86-66f8cb72486d" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.559886 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.560446 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nc9h65bh5fh9bhbch96h5cch548h5fdh57bh5c8h5b5h667h64h74h566h558h54bh64fh9fhd6h694h589h5c6h5d5h4hbfh577hdh5d4h7fhfcq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gxqmm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-54d5f75fb9-s2659_openstack(549beb55-bab4-4b1a-bf98-48de2e12956f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.565917 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-54d5f75fb9-s2659" podUID="549beb55-bab4-4b1a-bf98-48de2e12956f" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.567298 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.567455 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n89h589h676hfch59ch5c8h699h586h5bfh74hd5h54h599h586h56fh99h54dhch8fh668h678h5b6h694hcfh8ch564h654h66dh5dbh554h5c9h647q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jrsxt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7d578c747f-8f6hc_openstack(8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:22:32 crc kubenswrapper[4948]: E0220 08:22:32.570333 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7d578c747f-8f6hc" podUID="8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4" Feb 20 08:22:36 crc kubenswrapper[4948]: I0220 08:22:36.065516 4948 generic.go:334] "Generic (PLEG): container finished" podID="bb362d0e-68b4-4d48-885c-a6894ac71e6e" containerID="cb7241154041475439806eb631212532adb6e6775adf70e62bdf5cf4b6d19276" exitCode=0 Feb 20 08:22:36 crc kubenswrapper[4948]: I0220 08:22:36.065680 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-g5zxh" event={"ID":"bb362d0e-68b4-4d48-885c-a6894ac71e6e","Type":"ContainerDied","Data":"cb7241154041475439806eb631212532adb6e6775adf70e62bdf5cf4b6d19276"} Feb 20 08:22:39 crc kubenswrapper[4948]: I0220 08:22:39.849489 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-tkqzb" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: i/o timeout" Feb 20 08:22:40 crc kubenswrapper[4948]: E0220 08:22:40.010363 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Feb 20 08:22:40 crc kubenswrapper[4948]: E0220 08:22:40.010575 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gmlgz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-c7zzv_openstack(f1f108e1-27bb-433e-b7e1-f9cc7f778182): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:22:40 crc kubenswrapper[4948]: E0220 08:22:40.011719 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-c7zzv" podUID="f1f108e1-27bb-433e-b7e1-f9cc7f778182" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.079797 4948 scope.go:117] "RemoveContainer" containerID="fdfa9ea96d17b0fbac466add147012945b8f9c77bfe4c7e3939115948c20ff15" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.102156 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tkqzb" event={"ID":"da816a64-160f-4a5c-b945-b8fa467208d6","Type":"ContainerDied","Data":"32b2c84d98711b304188499ff3c33cd5bf7af6d44c7a701d021f0522c169184b"} Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.102421 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32b2c84d98711b304188499ff3c33cd5bf7af6d44c7a701d021f0522c169184b" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.110075 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e945937-1e9d-465b-a68f-c755119ab6f9","Type":"ContainerDied","Data":"feaec65599576490d6ae435b0c4ffac29904876fc82b7f96fd44c9df4357c090"} Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.110149 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="feaec65599576490d6ae435b0c4ffac29904876fc82b7f96fd44c9df4357c090" Feb 20 08:22:40 crc kubenswrapper[4948]: E0220 08:22:40.111209 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-c7zzv" podUID="f1f108e1-27bb-433e-b7e1-f9cc7f778182" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.225916 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.232393 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.237478 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.246885 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.262959 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.278197 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.408854 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-logs\") pod \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.408922 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vp59z\" (UniqueName: \"kubernetes.io/projected/136dd621-a5c5-481a-9b86-66f8cb72486d-kube-api-access-vp59z\") pod \"136dd621-a5c5-481a-9b86-66f8cb72486d\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409115 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/549beb55-bab4-4b1a-bf98-48de2e12956f-logs\") pod \"549beb55-bab4-4b1a-bf98-48de2e12956f\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409153 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svj9w\" (UniqueName: \"kubernetes.io/projected/6e945937-1e9d-465b-a68f-c755119ab6f9-kube-api-access-svj9w\") pod \"6e945937-1e9d-465b-a68f-c755119ab6f9\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409197 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wc2m\" (UniqueName: \"kubernetes.io/projected/da816a64-160f-4a5c-b945-b8fa467208d6-kube-api-access-4wc2m\") pod \"da816a64-160f-4a5c-b945-b8fa467208d6\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409257 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-scripts\") pod \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409312 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-scripts\") pod \"6e945937-1e9d-465b-a68f-c755119ab6f9\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409350 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-config-data\") pod \"6e945937-1e9d-465b-a68f-c755119ab6f9\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409382 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-config-data\") pod \"136dd621-a5c5-481a-9b86-66f8cb72486d\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409419 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-combined-ca-bundle\") pod \"6e945937-1e9d-465b-a68f-c755119ab6f9\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409462 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-config-data\") pod \"549beb55-bab4-4b1a-bf98-48de2e12956f\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409496 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-config-data\") pod \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409549 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-dns-svc\") pod \"da816a64-160f-4a5c-b945-b8fa467208d6\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409584 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npm75\" (UniqueName: \"kubernetes.io/projected/bb362d0e-68b4-4d48-885c-a6894ac71e6e-kube-api-access-npm75\") pod \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409637 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-horizon-secret-key\") pod \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409675 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-combined-ca-bundle\") pod \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409723 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/136dd621-a5c5-481a-9b86-66f8cb72486d-logs\") pod \"136dd621-a5c5-481a-9b86-66f8cb72486d\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409774 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrsxt\" (UniqueName: \"kubernetes.io/projected/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-kube-api-access-jrsxt\") pod \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\" (UID: \"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409805 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxqmm\" (UniqueName: \"kubernetes.io/projected/549beb55-bab4-4b1a-bf98-48de2e12956f-kube-api-access-gxqmm\") pod \"549beb55-bab4-4b1a-bf98-48de2e12956f\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409846 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-httpd-run\") pod \"6e945937-1e9d-465b-a68f-c755119ab6f9\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409878 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-sb\") pod \"da816a64-160f-4a5c-b945-b8fa467208d6\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409928 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-config\") pod \"da816a64-160f-4a5c-b945-b8fa467208d6\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.409994 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-nb\") pod \"da816a64-160f-4a5c-b945-b8fa467208d6\" (UID: \"da816a64-160f-4a5c-b945-b8fa467208d6\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.410035 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-config\") pod \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\" (UID: \"bb362d0e-68b4-4d48-885c-a6894ac71e6e\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.410065 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-logs\") pod \"6e945937-1e9d-465b-a68f-c755119ab6f9\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.410116 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-scripts\") pod \"549beb55-bab4-4b1a-bf98-48de2e12956f\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.410156 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-scripts\") pod \"136dd621-a5c5-481a-9b86-66f8cb72486d\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.410187 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/549beb55-bab4-4b1a-bf98-48de2e12956f-horizon-secret-key\") pod \"549beb55-bab4-4b1a-bf98-48de2e12956f\" (UID: \"549beb55-bab4-4b1a-bf98-48de2e12956f\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.410233 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/136dd621-a5c5-481a-9b86-66f8cb72486d-horizon-secret-key\") pod \"136dd621-a5c5-481a-9b86-66f8cb72486d\" (UID: \"136dd621-a5c5-481a-9b86-66f8cb72486d\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.410258 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"6e945937-1e9d-465b-a68f-c755119ab6f9\" (UID: \"6e945937-1e9d-465b-a68f-c755119ab6f9\") " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.411213 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-logs" (OuterVolumeSpecName: "logs") pod "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4" (UID: "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.411865 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-scripts" (OuterVolumeSpecName: "scripts") pod "549beb55-bab4-4b1a-bf98-48de2e12956f" (UID: "549beb55-bab4-4b1a-bf98-48de2e12956f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.411962 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-config-data" (OuterVolumeSpecName: "config-data") pod "549beb55-bab4-4b1a-bf98-48de2e12956f" (UID: "549beb55-bab4-4b1a-bf98-48de2e12956f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.412505 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-config-data" (OuterVolumeSpecName: "config-data") pod "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4" (UID: "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.413145 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/549beb55-bab4-4b1a-bf98-48de2e12956f-logs" (OuterVolumeSpecName: "logs") pod "549beb55-bab4-4b1a-bf98-48de2e12956f" (UID: "549beb55-bab4-4b1a-bf98-48de2e12956f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.412812 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6e945937-1e9d-465b-a68f-c755119ab6f9" (UID: "6e945937-1e9d-465b-a68f-c755119ab6f9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.413523 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/136dd621-a5c5-481a-9b86-66f8cb72486d-logs" (OuterVolumeSpecName: "logs") pod "136dd621-a5c5-481a-9b86-66f8cb72486d" (UID: "136dd621-a5c5-481a-9b86-66f8cb72486d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.413737 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-logs" (OuterVolumeSpecName: "logs") pod "6e945937-1e9d-465b-a68f-c755119ab6f9" (UID: "6e945937-1e9d-465b-a68f-c755119ab6f9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.413735 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-config-data" (OuterVolumeSpecName: "config-data") pod "136dd621-a5c5-481a-9b86-66f8cb72486d" (UID: "136dd621-a5c5-481a-9b86-66f8cb72486d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.413765 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-scripts" (OuterVolumeSpecName: "scripts") pod "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4" (UID: "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.413940 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-scripts" (OuterVolumeSpecName: "scripts") pod "136dd621-a5c5-481a-9b86-66f8cb72486d" (UID: "136dd621-a5c5-481a-9b86-66f8cb72486d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.419538 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/136dd621-a5c5-481a-9b86-66f8cb72486d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "136dd621-a5c5-481a-9b86-66f8cb72486d" (UID: "136dd621-a5c5-481a-9b86-66f8cb72486d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.419607 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e945937-1e9d-465b-a68f-c755119ab6f9-kube-api-access-svj9w" (OuterVolumeSpecName: "kube-api-access-svj9w") pod "6e945937-1e9d-465b-a68f-c755119ab6f9" (UID: "6e945937-1e9d-465b-a68f-c755119ab6f9"). InnerVolumeSpecName "kube-api-access-svj9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.420048 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb362d0e-68b4-4d48-885c-a6894ac71e6e-kube-api-access-npm75" (OuterVolumeSpecName: "kube-api-access-npm75") pod "bb362d0e-68b4-4d48-885c-a6894ac71e6e" (UID: "bb362d0e-68b4-4d48-885c-a6894ac71e6e"). InnerVolumeSpecName "kube-api-access-npm75". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.420422 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/549beb55-bab4-4b1a-bf98-48de2e12956f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "549beb55-bab4-4b1a-bf98-48de2e12956f" (UID: "549beb55-bab4-4b1a-bf98-48de2e12956f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.420445 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "6e945937-1e9d-465b-a68f-c755119ab6f9" (UID: "6e945937-1e9d-465b-a68f-c755119ab6f9"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.420510 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da816a64-160f-4a5c-b945-b8fa467208d6-kube-api-access-4wc2m" (OuterVolumeSpecName: "kube-api-access-4wc2m") pod "da816a64-160f-4a5c-b945-b8fa467208d6" (UID: "da816a64-160f-4a5c-b945-b8fa467208d6"). InnerVolumeSpecName "kube-api-access-4wc2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.420542 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/136dd621-a5c5-481a-9b86-66f8cb72486d-kube-api-access-vp59z" (OuterVolumeSpecName: "kube-api-access-vp59z") pod "136dd621-a5c5-481a-9b86-66f8cb72486d" (UID: "136dd621-a5c5-481a-9b86-66f8cb72486d"). InnerVolumeSpecName "kube-api-access-vp59z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.422311 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4" (UID: "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.422753 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-scripts" (OuterVolumeSpecName: "scripts") pod "6e945937-1e9d-465b-a68f-c755119ab6f9" (UID: "6e945937-1e9d-465b-a68f-c755119ab6f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.423566 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-kube-api-access-jrsxt" (OuterVolumeSpecName: "kube-api-access-jrsxt") pod "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4" (UID: "8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4"). InnerVolumeSpecName "kube-api-access-jrsxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.426712 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/549beb55-bab4-4b1a-bf98-48de2e12956f-kube-api-access-gxqmm" (OuterVolumeSpecName: "kube-api-access-gxqmm") pod "549beb55-bab4-4b1a-bf98-48de2e12956f" (UID: "549beb55-bab4-4b1a-bf98-48de2e12956f"). InnerVolumeSpecName "kube-api-access-gxqmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.446992 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-config" (OuterVolumeSpecName: "config") pod "bb362d0e-68b4-4d48-885c-a6894ac71e6e" (UID: "bb362d0e-68b4-4d48-885c-a6894ac71e6e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.458817 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e945937-1e9d-465b-a68f-c755119ab6f9" (UID: "6e945937-1e9d-465b-a68f-c755119ab6f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.459342 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb362d0e-68b4-4d48-885c-a6894ac71e6e" (UID: "bb362d0e-68b4-4d48-885c-a6894ac71e6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.472920 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-config" (OuterVolumeSpecName: "config") pod "da816a64-160f-4a5c-b945-b8fa467208d6" (UID: "da816a64-160f-4a5c-b945-b8fa467208d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.483283 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "da816a64-160f-4a5c-b945-b8fa467208d6" (UID: "da816a64-160f-4a5c-b945-b8fa467208d6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.485653 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "da816a64-160f-4a5c-b945-b8fa467208d6" (UID: "da816a64-160f-4a5c-b945-b8fa467208d6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.485646 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-config-data" (OuterVolumeSpecName: "config-data") pod "6e945937-1e9d-465b-a68f-c755119ab6f9" (UID: "6e945937-1e9d-465b-a68f-c755119ab6f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.486779 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "da816a64-160f-4a5c-b945-b8fa467208d6" (UID: "da816a64-160f-4a5c-b945-b8fa467208d6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.511895 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/136dd621-a5c5-481a-9b86-66f8cb72486d-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.511926 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrsxt\" (UniqueName: \"kubernetes.io/projected/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-kube-api-access-jrsxt\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.511939 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxqmm\" (UniqueName: \"kubernetes.io/projected/549beb55-bab4-4b1a-bf98-48de2e12956f-kube-api-access-gxqmm\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.511948 4948 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.511956 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.511965 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.511984 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.511993 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512001 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e945937-1e9d-465b-a68f-c755119ab6f9-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512010 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512019 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512026 4948 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/549beb55-bab4-4b1a-bf98-48de2e12956f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512034 4948 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/136dd621-a5c5-481a-9b86-66f8cb72486d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512066 4948 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512076 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512084 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vp59z\" (UniqueName: \"kubernetes.io/projected/136dd621-a5c5-481a-9b86-66f8cb72486d-kube-api-access-vp59z\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512091 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/549beb55-bab4-4b1a-bf98-48de2e12956f-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512099 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svj9w\" (UniqueName: \"kubernetes.io/projected/6e945937-1e9d-465b-a68f-c755119ab6f9-kube-api-access-svj9w\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512107 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wc2m\" (UniqueName: \"kubernetes.io/projected/da816a64-160f-4a5c-b945-b8fa467208d6-kube-api-access-4wc2m\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512115 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512122 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512129 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512139 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/136dd621-a5c5-481a-9b86-66f8cb72486d-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512148 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e945937-1e9d-465b-a68f-c755119ab6f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512156 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/549beb55-bab4-4b1a-bf98-48de2e12956f-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512164 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512172 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da816a64-160f-4a5c-b945-b8fa467208d6-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512180 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npm75\" (UniqueName: \"kubernetes.io/projected/bb362d0e-68b4-4d48-885c-a6894ac71e6e-kube-api-access-npm75\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512188 4948 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.512197 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb362d0e-68b4-4d48-885c-a6894ac71e6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.532293 4948 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Feb 20 08:22:40 crc kubenswrapper[4948]: I0220 08:22:40.614374 4948 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.125909 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-g5zxh" event={"ID":"bb362d0e-68b4-4d48-885c-a6894ac71e6e","Type":"ContainerDied","Data":"fb0ac2219219fe66e5a42c9ab8b34f9c1f674fc235bce862e34dedbfa83dd480"} Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.125937 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-g5zxh" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.125953 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb0ac2219219fe66e5a42c9ab8b34f9c1f674fc235bce862e34dedbfa83dd480" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.127790 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c9c4f5fc-2vtb4" event={"ID":"136dd621-a5c5-481a-9b86-66f8cb72486d","Type":"ContainerDied","Data":"45d6073bbc41d12c6fd4a83a0736dfc692eb915e2f643d37fcfcdd4574770ee8"} Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.128272 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68c9c4f5fc-2vtb4" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.133310 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d578c747f-8f6hc" event={"ID":"8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4","Type":"ContainerDied","Data":"3f991d1b028cfdcb1512441c7603f82fee2c2bba86ea276855760a70987a1d8f"} Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.133357 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d578c747f-8f6hc" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.138509 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.138605 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54d5f75fb9-s2659" event={"ID":"549beb55-bab4-4b1a-bf98-48de2e12956f","Type":"ContainerDied","Data":"27986056f94b762ad4c0ec91cfd1562122b907c3b47800656945a8588571afe8"} Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.138669 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54d5f75fb9-s2659" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.143086 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tkqzb" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.222853 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68c9c4f5fc-2vtb4"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.232860 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-68c9c4f5fc-2vtb4"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.251942 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tkqzb"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.264702 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tkqzb"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.286327 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-54d5f75fb9-s2659"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.295382 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-54d5f75fb9-s2659"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.313775 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d578c747f-8f6hc"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.323245 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7d578c747f-8f6hc"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.332958 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.342088 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.352176 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:41 crc kubenswrapper[4948]: E0220 08:22:41.352611 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="dnsmasq-dns" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.352632 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="dnsmasq-dns" Feb 20 08:22:41 crc kubenswrapper[4948]: E0220 08:22:41.352659 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerName="glance-httpd" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.352667 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerName="glance-httpd" Feb 20 08:22:41 crc kubenswrapper[4948]: E0220 08:22:41.352689 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb362d0e-68b4-4d48-885c-a6894ac71e6e" containerName="neutron-db-sync" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.352698 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb362d0e-68b4-4d48-885c-a6894ac71e6e" containerName="neutron-db-sync" Feb 20 08:22:41 crc kubenswrapper[4948]: E0220 08:22:41.352716 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="init" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.352723 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="init" Feb 20 08:22:41 crc kubenswrapper[4948]: E0220 08:22:41.352738 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerName="glance-log" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.352745 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerName="glance-log" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.352951 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerName="glance-log" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.352990 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" containerName="glance-httpd" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.353001 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb362d0e-68b4-4d48-885c-a6894ac71e6e" containerName="neutron-db-sync" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.353010 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="dnsmasq-dns" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.354129 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.356114 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.356513 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.359628 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.472043 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-v8n5p"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.473773 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.494440 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-v8n5p"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.527754 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6phw\" (UniqueName: \"kubernetes.io/projected/33b35605-1057-4e64-99a7-a7273351d6c2-kube-api-access-q6phw\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.527818 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.527876 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.527911 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.527939 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.527999 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.528028 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-logs\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.528079 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.581165 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b5c5cc45d-7dcbq"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.582409 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.584808 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.585199 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.585573 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4txrz" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.585949 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.603461 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b5c5cc45d-7dcbq"] Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630657 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgbxd\" (UniqueName: \"kubernetes.io/projected/0cc4ba01-323f-4bcb-9bc3-d639514f6088-kube-api-access-sgbxd\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630705 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630738 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-svc\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630773 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6phw\" (UniqueName: \"kubernetes.io/projected/33b35605-1057-4e64-99a7-a7273351d6c2-kube-api-access-q6phw\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630798 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630854 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630884 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630902 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630927 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-config\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.630961 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.631001 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-logs\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.631030 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.631064 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.631093 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.631722 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.632839 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.632879 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-logs\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.636120 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.638286 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.639498 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.649405 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.650954 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6phw\" (UniqueName: \"kubernetes.io/projected/33b35605-1057-4e64-99a7-a7273351d6c2-kube-api-access-q6phw\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.679025 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.733492 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="136dd621-a5c5-481a-9b86-66f8cb72486d" path="/var/lib/kubelet/pods/136dd621-a5c5-481a-9b86-66f8cb72486d/volumes" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.733929 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="549beb55-bab4-4b1a-bf98-48de2e12956f" path="/var/lib/kubelet/pods/549beb55-bab4-4b1a-bf98-48de2e12956f/volumes" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.734399 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e945937-1e9d-465b-a68f-c755119ab6f9" path="/var/lib/kubelet/pods/6e945937-1e9d-465b-a68f-c755119ab6f9/volumes" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.734881 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-ovndb-tls-certs\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.734952 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735004 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khls2\" (UniqueName: \"kubernetes.io/projected/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-kube-api-access-khls2\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735039 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735099 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-config\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735129 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-httpd-config\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735153 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgbxd\" (UniqueName: \"kubernetes.io/projected/0cc4ba01-323f-4bcb-9bc3-d639514f6088-kube-api-access-sgbxd\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735181 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735228 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-svc\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735282 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-combined-ca-bundle\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.735317 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-config\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.736137 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4" path="/var/lib/kubelet/pods/8425fd12-ae45-4e7e-af4d-9bfc3c3d0eb4/volumes" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.736548 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" path="/var/lib/kubelet/pods/da816a64-160f-4a5c-b945-b8fa467208d6/volumes" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.736569 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.736631 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.737212 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-config\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.737530 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-svc\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.737678 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.754051 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgbxd\" (UniqueName: \"kubernetes.io/projected/0cc4ba01-323f-4bcb-9bc3-d639514f6088-kube-api-access-sgbxd\") pod \"dnsmasq-dns-55f844cf75-v8n5p\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.810650 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.837142 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khls2\" (UniqueName: \"kubernetes.io/projected/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-kube-api-access-khls2\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.837225 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-config\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.837262 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-httpd-config\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.837387 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-combined-ca-bundle\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.837439 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-ovndb-tls-certs\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.843164 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-ovndb-tls-certs\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.844152 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-httpd-config\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.847818 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-config\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.852685 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-combined-ca-bundle\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.859647 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khls2\" (UniqueName: \"kubernetes.io/projected/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-kube-api-access-khls2\") pod \"neutron-b5c5cc45d-7dcbq\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.908649 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:41 crc kubenswrapper[4948]: I0220 08:22:41.978604 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:42 crc kubenswrapper[4948]: E0220 08:22:42.059487 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Feb 20 08:22:42 crc kubenswrapper[4948]: E0220 08:22:42.059678 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fwtv5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-xt2zq_openstack(f706b741-86f9-49a3-95d5-85ee62eb3668): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:22:42 crc kubenswrapper[4948]: E0220 08:22:42.060815 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-xt2zq" podUID="f706b741-86f9-49a3-95d5-85ee62eb3668" Feb 20 08:22:42 crc kubenswrapper[4948]: E0220 08:22:42.177132 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-xt2zq" podUID="f706b741-86f9-49a3-95d5-85ee62eb3668" Feb 20 08:22:42 crc kubenswrapper[4948]: W0220 08:22:42.675476 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b9a072a_a040_414c_96c7_cf1454148745.slice/crio-6dcfb76c7284437f090e639d57144c71edd62b4bab7242fd64ad46e096ae5a7b WatchSource:0}: Error finding container 6dcfb76c7284437f090e639d57144c71edd62b4bab7242fd64ad46e096ae5a7b: Status 404 returned error can't find the container with id 6dcfb76c7284437f090e639d57144c71edd62b4bab7242fd64ad46e096ae5a7b Feb 20 08:22:42 crc kubenswrapper[4948]: I0220 08:22:42.680594 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-685c4db87b-rw57h"] Feb 20 08:22:42 crc kubenswrapper[4948]: I0220 08:22:42.772116 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g27hk"] Feb 20 08:22:42 crc kubenswrapper[4948]: I0220 08:22:42.797821 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66b4d4cb8-hgmkd"] Feb 20 08:22:42 crc kubenswrapper[4948]: I0220 08:22:42.842062 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:42 crc kubenswrapper[4948]: W0220 08:22:42.852227 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0f82f51_8a20_402d_8861_6c57a4e5af0d.slice/crio-70cd0cac387b5f659d055b8820f12fcea594407eebec35ab497b573fb1f4db5a WatchSource:0}: Error finding container 70cd0cac387b5f659d055b8820f12fcea594407eebec35ab497b573fb1f4db5a: Status 404 returned error can't find the container with id 70cd0cac387b5f659d055b8820f12fcea594407eebec35ab497b573fb1f4db5a Feb 20 08:22:42 crc kubenswrapper[4948]: I0220 08:22:42.883468 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hh5s2"] Feb 20 08:22:42 crc kubenswrapper[4948]: W0220 08:22:42.898317 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd84a71b6_7503_4150_a5c4_4579b08a669a.slice/crio-21944c755e2a58a194f7e12be68919b81d0c021ee6380fb185cda3cec0fcee37 WatchSource:0}: Error finding container 21944c755e2a58a194f7e12be68919b81d0c021ee6380fb185cda3cec0fcee37: Status 404 returned error can't find the container with id 21944c755e2a58a194f7e12be68919b81d0c021ee6380fb185cda3cec0fcee37 Feb 20 08:22:42 crc kubenswrapper[4948]: I0220 08:22:42.903703 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v29fx"] Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.029088 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-v8n5p"] Feb 20 08:22:43 crc kubenswrapper[4948]: W0220 08:22:43.048032 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cc4ba01_323f_4bcb_9bc3_d639514f6088.slice/crio-5ba21a5e281f6be663a46ae25fc6e9402ee9bcd0e6caec5c0c3b4658222b76c1 WatchSource:0}: Error finding container 5ba21a5e281f6be663a46ae25fc6e9402ee9bcd0e6caec5c0c3b4658222b76c1: Status 404 returned error can't find the container with id 5ba21a5e281f6be663a46ae25fc6e9402ee9bcd0e6caec5c0c3b4658222b76c1 Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.187480 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6mz7x" event={"ID":"1227b752-8a3c-4c01-91bf-0662c1b1a231","Type":"ContainerStarted","Data":"b8d5cd74c5cbd29dae34c7b1cedbde475f451d2de997a59b511db6cc61ba0a29"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.203114 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66b4d4cb8-hgmkd" event={"ID":"d1046503-c6d3-49bf-9523-2f897fcb82ba","Type":"ContainerStarted","Data":"0fc9151c9d2a99b81f2c9346beebf0499347807bacbd7c95da55982c7422c38e"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.224417 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.227846 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-6mz7x" podStartSLOduration=5.800811674 podStartE2EDuration="29.2278275s" podCreationTimestamp="2026-02-20 08:22:14 +0000 UTC" firstStartedPulling="2026-02-20 08:22:16.65605074 +0000 UTC m=+985.630545560" lastFinishedPulling="2026-02-20 08:22:40.083066546 +0000 UTC m=+1009.057561386" observedRunningTime="2026-02-20 08:22:43.224868687 +0000 UTC m=+1012.199363507" watchObservedRunningTime="2026-02-20 08:22:43.2278275 +0000 UTC m=+1012.202322320" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.244168 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerStarted","Data":"f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.259945 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6729921-0a12-46a0-824f-810df88ae426" containerID="c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a" exitCode=0 Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.260031 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g27hk" event={"ID":"d6729921-0a12-46a0-824f-810df88ae426","Type":"ContainerDied","Data":"c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.260060 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g27hk" event={"ID":"d6729921-0a12-46a0-824f-810df88ae426","Type":"ContainerStarted","Data":"b7570c60dc7ea9b2dda7deeb7d6ef87e7350a05f972449d37dfa272d75ce2941"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.272664 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh5s2" event={"ID":"426b37f9-4acb-4d28-a55c-3fcde2600002","Type":"ContainerStarted","Data":"6c8a403f4e4cb34b74e5da72f8bdf18b9a2a922052dce8500240a25dba684641"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.302253 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v29fx" event={"ID":"d84a71b6-7503-4150-a5c4-4579b08a669a","Type":"ContainerStarted","Data":"21944c755e2a58a194f7e12be68919b81d0c021ee6380fb185cda3cec0fcee37"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.307192 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e0f82f51-8a20-402d-8861-6c57a4e5af0d","Type":"ContainerStarted","Data":"70cd0cac387b5f659d055b8820f12fcea594407eebec35ab497b573fb1f4db5a"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.314320 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" event={"ID":"0cc4ba01-323f-4bcb-9bc3-d639514f6088","Type":"ContainerStarted","Data":"5ba21a5e281f6be663a46ae25fc6e9402ee9bcd0e6caec5c0c3b4658222b76c1"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.320863 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-685c4db87b-rw57h" event={"ID":"7b9a072a-a040-414c-96c7-cf1454148745","Type":"ContainerStarted","Data":"6dcfb76c7284437f090e639d57144c71edd62b4bab7242fd64ad46e096ae5a7b"} Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.548894 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-685bb7f5c-jd9nm"] Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.556874 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.564932 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.566202 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.575810 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-685bb7f5c-jd9nm"] Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.705591 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh687\" (UniqueName: \"kubernetes.io/projected/e0661780-ca2b-42ae-be25-5d0e914fbb49-kube-api-access-fh687\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.706202 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-combined-ca-bundle\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.706360 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-public-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.706382 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-ovndb-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.706909 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-config\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.706950 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-internal-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.707630 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-httpd-config\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.808739 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-combined-ca-bundle\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.808812 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-public-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.808832 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-ovndb-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.810208 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-config\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.810328 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-internal-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.810608 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-httpd-config\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.810633 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh687\" (UniqueName: \"kubernetes.io/projected/e0661780-ca2b-42ae-be25-5d0e914fbb49-kube-api-access-fh687\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.812646 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-public-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.812912 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-combined-ca-bundle\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.817554 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-ovndb-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.818235 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-config\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.825815 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-internal-tls-certs\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.826512 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-httpd-config\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.830687 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh687\" (UniqueName: \"kubernetes.io/projected/e0661780-ca2b-42ae-be25-5d0e914fbb49-kube-api-access-fh687\") pod \"neutron-685bb7f5c-jd9nm\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:43 crc kubenswrapper[4948]: I0220 08:22:43.923624 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.076452 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b5c5cc45d-7dcbq"] Feb 20 08:22:44 crc kubenswrapper[4948]: W0220 08:22:44.114685 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd1b85e2_1b5f_4993_bc66_72daaf6aa79d.slice/crio-4c4522a0ae94ffaed4b0ddc1a755c0498d0a11f95af5418408b70c0351708567 WatchSource:0}: Error finding container 4c4522a0ae94ffaed4b0ddc1a755c0498d0a11f95af5418408b70c0351708567: Status 404 returned error can't find the container with id 4c4522a0ae94ffaed4b0ddc1a755c0498d0a11f95af5418408b70c0351708567 Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.344115 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-685c4db87b-rw57h" event={"ID":"7b9a072a-a040-414c-96c7-cf1454148745","Type":"ContainerStarted","Data":"827809de35c30bbe5e9607a26506e33891d08d97441e79c25a85c226aa503507"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.345444 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-685c4db87b-rw57h" event={"ID":"7b9a072a-a040-414c-96c7-cf1454148745","Type":"ContainerStarted","Data":"f855ae3e2aaeae487e0cde870ef0f634ece3468bf55857211d1d8d871cc6a217"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.348308 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e0f82f51-8a20-402d-8861-6c57a4e5af0d","Type":"ContainerStarted","Data":"7020629c426e657a58b5859f74b5d55f5c4fb964822922a0eb88d95415abe7d8"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.354251 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66b4d4cb8-hgmkd" event={"ID":"d1046503-c6d3-49bf-9523-2f897fcb82ba","Type":"ContainerStarted","Data":"577865ca69f0d8457d617948e5cac2be0a124e4620edc60df1395130de8a2334"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.357391 4948 generic.go:334] "Generic (PLEG): container finished" podID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerID="7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743" exitCode=0 Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.357826 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" event={"ID":"0cc4ba01-323f-4bcb-9bc3-d639514f6088","Type":"ContainerDied","Data":"7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.364514 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5c5cc45d-7dcbq" event={"ID":"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d","Type":"ContainerStarted","Data":"4c4522a0ae94ffaed4b0ddc1a755c0498d0a11f95af5418408b70c0351708567"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.370455 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-685c4db87b-rw57h" podStartSLOduration=18.836466751 podStartE2EDuration="19.370432985s" podCreationTimestamp="2026-02-20 08:22:25 +0000 UTC" firstStartedPulling="2026-02-20 08:22:42.677224764 +0000 UTC m=+1011.651719584" lastFinishedPulling="2026-02-20 08:22:43.211190998 +0000 UTC m=+1012.185685818" observedRunningTime="2026-02-20 08:22:44.364765685 +0000 UTC m=+1013.339260505" watchObservedRunningTime="2026-02-20 08:22:44.370432985 +0000 UTC m=+1013.344927805" Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.371291 4948 generic.go:334] "Generic (PLEG): container finished" podID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerID="62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863" exitCode=0 Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.371373 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh5s2" event={"ID":"426b37f9-4acb-4d28-a55c-3fcde2600002","Type":"ContainerDied","Data":"62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.373792 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v29fx" event={"ID":"d84a71b6-7503-4150-a5c4-4579b08a669a","Type":"ContainerStarted","Data":"87c2cdb1a2b7f5c38a01dd7f884ce5938563a33612ebaf51f20ceac5532f97f2"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.394335 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33b35605-1057-4e64-99a7-a7273351d6c2","Type":"ContainerStarted","Data":"9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.394385 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33b35605-1057-4e64-99a7-a7273351d6c2","Type":"ContainerStarted","Data":"3abd35d213010021a69fde90dd757e9181aec2b84f4e7f71376ea0edef4300a5"} Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.412173 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-v29fx" podStartSLOduration=16.412158248 podStartE2EDuration="16.412158248s" podCreationTimestamp="2026-02-20 08:22:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:44.408938268 +0000 UTC m=+1013.383433108" watchObservedRunningTime="2026-02-20 08:22:44.412158248 +0000 UTC m=+1013.386653068" Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.631493 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-685bb7f5c-jd9nm"] Feb 20 08:22:44 crc kubenswrapper[4948]: I0220 08:22:44.851277 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-tkqzb" podUID="da816a64-160f-4a5c-b945-b8fa467208d6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: i/o timeout" Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.404777 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685bb7f5c-jd9nm" event={"ID":"e0661780-ca2b-42ae-be25-5d0e914fbb49","Type":"ContainerStarted","Data":"4885fc2e4d8fe9fc0c29512f6a5fd4f1b7a87efbf9c10fe17b8bf4fe7f1ca02a"} Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.407096 4948 generic.go:334] "Generic (PLEG): container finished" podID="1227b752-8a3c-4c01-91bf-0662c1b1a231" containerID="b8d5cd74c5cbd29dae34c7b1cedbde475f451d2de997a59b511db6cc61ba0a29" exitCode=0 Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.407156 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6mz7x" event={"ID":"1227b752-8a3c-4c01-91bf-0662c1b1a231","Type":"ContainerDied","Data":"b8d5cd74c5cbd29dae34c7b1cedbde475f451d2de997a59b511db6cc61ba0a29"} Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.409489 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66b4d4cb8-hgmkd" event={"ID":"d1046503-c6d3-49bf-9523-2f897fcb82ba","Type":"ContainerStarted","Data":"a1b60a312d88b84db5225c1aff5290457499b2775b6fffec7ca3246896b28523"} Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.413413 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" event={"ID":"0cc4ba01-323f-4bcb-9bc3-d639514f6088","Type":"ContainerStarted","Data":"ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386"} Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.413575 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.415048 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5c5cc45d-7dcbq" event={"ID":"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d","Type":"ContainerStarted","Data":"fceb35ea8befbc40c2edb25ea5ef78a9c4a759e83b13e451735a1e2ccf6a99a8"} Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.418263 4948 generic.go:334] "Generic (PLEG): container finished" podID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerID="2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d" exitCode=0 Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.418446 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh5s2" event={"ID":"426b37f9-4acb-4d28-a55c-3fcde2600002","Type":"ContainerDied","Data":"2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d"} Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.450362 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" podStartSLOduration=4.45034643 podStartE2EDuration="4.45034643s" podCreationTimestamp="2026-02-20 08:22:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:45.4475054 +0000 UTC m=+1014.422000220" watchObservedRunningTime="2026-02-20 08:22:45.45034643 +0000 UTC m=+1014.424841250" Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.477270 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.477314 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.498401 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-66b4d4cb8-hgmkd" podStartSLOduration=19.956081589 podStartE2EDuration="20.498382519s" podCreationTimestamp="2026-02-20 08:22:25 +0000 UTC" firstStartedPulling="2026-02-20 08:22:42.822392586 +0000 UTC m=+1011.796887406" lastFinishedPulling="2026-02-20 08:22:43.364693516 +0000 UTC m=+1012.339188336" observedRunningTime="2026-02-20 08:22:45.492868092 +0000 UTC m=+1014.467362922" watchObservedRunningTime="2026-02-20 08:22:45.498382519 +0000 UTC m=+1014.472877339" Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.643332 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:45 crc kubenswrapper[4948]: I0220 08:22:45.643510 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.874570 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.976446 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-config-data\") pod \"1227b752-8a3c-4c01-91bf-0662c1b1a231\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.976536 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1227b752-8a3c-4c01-91bf-0662c1b1a231-logs\") pod \"1227b752-8a3c-4c01-91bf-0662c1b1a231\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.976630 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rtp8\" (UniqueName: \"kubernetes.io/projected/1227b752-8a3c-4c01-91bf-0662c1b1a231-kube-api-access-5rtp8\") pod \"1227b752-8a3c-4c01-91bf-0662c1b1a231\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.976683 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-scripts\") pod \"1227b752-8a3c-4c01-91bf-0662c1b1a231\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.976805 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-combined-ca-bundle\") pod \"1227b752-8a3c-4c01-91bf-0662c1b1a231\" (UID: \"1227b752-8a3c-4c01-91bf-0662c1b1a231\") " Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.976933 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1227b752-8a3c-4c01-91bf-0662c1b1a231-logs" (OuterVolumeSpecName: "logs") pod "1227b752-8a3c-4c01-91bf-0662c1b1a231" (UID: "1227b752-8a3c-4c01-91bf-0662c1b1a231"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.977273 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1227b752-8a3c-4c01-91bf-0662c1b1a231-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.987392 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1227b752-8a3c-4c01-91bf-0662c1b1a231-kube-api-access-5rtp8" (OuterVolumeSpecName: "kube-api-access-5rtp8") pod "1227b752-8a3c-4c01-91bf-0662c1b1a231" (UID: "1227b752-8a3c-4c01-91bf-0662c1b1a231"). InnerVolumeSpecName "kube-api-access-5rtp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:46 crc kubenswrapper[4948]: I0220 08:22:46.992302 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-scripts" (OuterVolumeSpecName: "scripts") pod "1227b752-8a3c-4c01-91bf-0662c1b1a231" (UID: "1227b752-8a3c-4c01-91bf-0662c1b1a231"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.018304 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-config-data" (OuterVolumeSpecName: "config-data") pod "1227b752-8a3c-4c01-91bf-0662c1b1a231" (UID: "1227b752-8a3c-4c01-91bf-0662c1b1a231"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.039744 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1227b752-8a3c-4c01-91bf-0662c1b1a231" (UID: "1227b752-8a3c-4c01-91bf-0662c1b1a231"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.078448 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.078481 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.078495 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rtp8\" (UniqueName: \"kubernetes.io/projected/1227b752-8a3c-4c01-91bf-0662c1b1a231-kube-api-access-5rtp8\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.078507 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1227b752-8a3c-4c01-91bf-0662c1b1a231-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.484665 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6mz7x" event={"ID":"1227b752-8a3c-4c01-91bf-0662c1b1a231","Type":"ContainerDied","Data":"5b0f5c257994604a44a9cf0e2c06abc610dcd1eb5c3d34cfd415bfd74671d08f"} Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.484708 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b0f5c257994604a44a9cf0e2c06abc610dcd1eb5c3d34cfd415bfd74671d08f" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.484793 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6mz7x" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.487479 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerStarted","Data":"68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493"} Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.496895 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g27hk" event={"ID":"d6729921-0a12-46a0-824f-810df88ae426","Type":"ContainerStarted","Data":"bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793"} Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.563451 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-d46988dcb-rn6qs"] Feb 20 08:22:47 crc kubenswrapper[4948]: E0220 08:22:47.563850 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1227b752-8a3c-4c01-91bf-0662c1b1a231" containerName="placement-db-sync" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.563868 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1227b752-8a3c-4c01-91bf-0662c1b1a231" containerName="placement-db-sync" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.564073 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="1227b752-8a3c-4c01-91bf-0662c1b1a231" containerName="placement-db-sync" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.572666 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.584149 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.584301 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.584585 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.584838 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-2fq52" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.585220 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.588494 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d46988dcb-rn6qs"] Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.696404 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-internal-tls-certs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.696904 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bpmf\" (UniqueName: \"kubernetes.io/projected/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-kube-api-access-2bpmf\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.696951 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-public-tls-certs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.697021 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-combined-ca-bundle\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.697892 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-config-data\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.698009 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-scripts\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.698227 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-logs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.799570 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-config-data\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.799621 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-scripts\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.799642 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-logs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.799685 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-internal-tls-certs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.799740 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bpmf\" (UniqueName: \"kubernetes.io/projected/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-kube-api-access-2bpmf\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.799763 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-public-tls-certs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.799778 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-combined-ca-bundle\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.800184 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-logs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.811381 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-internal-tls-certs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.815595 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-public-tls-certs\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.815739 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-combined-ca-bundle\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.815893 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-config-data\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.817271 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-scripts\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.820798 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bpmf\" (UniqueName: \"kubernetes.io/projected/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-kube-api-access-2bpmf\") pod \"placement-d46988dcb-rn6qs\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:47 crc kubenswrapper[4948]: I0220 08:22:47.891354 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.507029 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33b35605-1057-4e64-99a7-a7273351d6c2","Type":"ContainerStarted","Data":"08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679"} Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.511199 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e0f82f51-8a20-402d-8861-6c57a4e5af0d","Type":"ContainerStarted","Data":"1a5b270055c8b2b5a29d54a228d2fd2e9bdac72aaebc130807e4057de5afdb1a"} Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.511348 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerName="glance-log" containerID="cri-o://7020629c426e657a58b5859f74b5d55f5c4fb964822922a0eb88d95415abe7d8" gracePeriod=30 Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.511409 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerName="glance-httpd" containerID="cri-o://1a5b270055c8b2b5a29d54a228d2fd2e9bdac72aaebc130807e4057de5afdb1a" gracePeriod=30 Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.520894 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5c5cc45d-7dcbq" event={"ID":"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d","Type":"ContainerStarted","Data":"4fc7cff150a17fca1672ea7d9eae41c85a175b0d1f41224639307a4d7025cd12"} Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.521116 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.524605 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6729921-0a12-46a0-824f-810df88ae426" containerID="bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793" exitCode=0 Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.524728 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g27hk" event={"ID":"d6729921-0a12-46a0-824f-810df88ae426","Type":"ContainerDied","Data":"bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793"} Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.525924 4948 generic.go:334] "Generic (PLEG): container finished" podID="d84a71b6-7503-4150-a5c4-4579b08a669a" containerID="87c2cdb1a2b7f5c38a01dd7f884ce5938563a33612ebaf51f20ceac5532f97f2" exitCode=0 Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.526055 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v29fx" event={"ID":"d84a71b6-7503-4150-a5c4-4579b08a669a","Type":"ContainerDied","Data":"87c2cdb1a2b7f5c38a01dd7f884ce5938563a33612ebaf51f20ceac5532f97f2"} Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.534953 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.534935674 podStartE2EDuration="7.534935674s" podCreationTimestamp="2026-02-20 08:22:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:48.530491234 +0000 UTC m=+1017.504986054" watchObservedRunningTime="2026-02-20 08:22:48.534935674 +0000 UTC m=+1017.509430494" Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.536594 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685bb7f5c-jd9nm" event={"ID":"e0661780-ca2b-42ae-be25-5d0e914fbb49","Type":"ContainerStarted","Data":"1ad25f5bfef4b0fa06a939cc9fb8e5bfa3dad488d3265a3cd5fa655c097dda5b"} Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.583011 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=26.582960403 podStartE2EDuration="26.582960403s" podCreationTimestamp="2026-02-20 08:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:48.571639442 +0000 UTC m=+1017.546134262" watchObservedRunningTime="2026-02-20 08:22:48.582960403 +0000 UTC m=+1017.557455223" Feb 20 08:22:48 crc kubenswrapper[4948]: I0220 08:22:48.612133 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-b5c5cc45d-7dcbq" podStartSLOduration=7.612108314 podStartE2EDuration="7.612108314s" podCreationTimestamp="2026-02-20 08:22:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:48.599041891 +0000 UTC m=+1017.573536711" watchObservedRunningTime="2026-02-20 08:22:48.612108314 +0000 UTC m=+1017.586603144" Feb 20 08:22:49 crc kubenswrapper[4948]: I0220 08:22:49.335448 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d46988dcb-rn6qs"] Feb 20 08:22:49 crc kubenswrapper[4948]: W0220 08:22:49.344491 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b1ee525_1a6e_461d_aa4d_178ab601ec9d.slice/crio-101e3a4bd62000e7fe0c14babf5f059dc5dad0bc38ccadaa7bd0301c2ad9ec7f WatchSource:0}: Error finding container 101e3a4bd62000e7fe0c14babf5f059dc5dad0bc38ccadaa7bd0301c2ad9ec7f: Status 404 returned error can't find the container with id 101e3a4bd62000e7fe0c14babf5f059dc5dad0bc38ccadaa7bd0301c2ad9ec7f Feb 20 08:22:49 crc kubenswrapper[4948]: I0220 08:22:49.552588 4948 generic.go:334] "Generic (PLEG): container finished" podID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerID="1a5b270055c8b2b5a29d54a228d2fd2e9bdac72aaebc130807e4057de5afdb1a" exitCode=0 Feb 20 08:22:49 crc kubenswrapper[4948]: I0220 08:22:49.553021 4948 generic.go:334] "Generic (PLEG): container finished" podID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerID="7020629c426e657a58b5859f74b5d55f5c4fb964822922a0eb88d95415abe7d8" exitCode=143 Feb 20 08:22:49 crc kubenswrapper[4948]: I0220 08:22:49.553078 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e0f82f51-8a20-402d-8861-6c57a4e5af0d","Type":"ContainerDied","Data":"1a5b270055c8b2b5a29d54a228d2fd2e9bdac72aaebc130807e4057de5afdb1a"} Feb 20 08:22:49 crc kubenswrapper[4948]: I0220 08:22:49.553111 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e0f82f51-8a20-402d-8861-6c57a4e5af0d","Type":"ContainerDied","Data":"7020629c426e657a58b5859f74b5d55f5c4fb964822922a0eb88d95415abe7d8"} Feb 20 08:22:49 crc kubenswrapper[4948]: I0220 08:22:49.556805 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d46988dcb-rn6qs" event={"ID":"8b1ee525-1a6e-461d-aa4d-178ab601ec9d","Type":"ContainerStarted","Data":"101e3a4bd62000e7fe0c14babf5f059dc5dad0bc38ccadaa7bd0301c2ad9ec7f"} Feb 20 08:22:49 crc kubenswrapper[4948]: I0220 08:22:49.562191 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685bb7f5c-jd9nm" event={"ID":"e0661780-ca2b-42ae-be25-5d0e914fbb49","Type":"ContainerStarted","Data":"d2ca1451e75be7daafc6907a8305baa229ed6441f1f79895bd2df00e5c17bfa7"} Feb 20 08:22:49 crc kubenswrapper[4948]: I0220 08:22:49.584445 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-685bb7f5c-jd9nm" podStartSLOduration=6.584426876 podStartE2EDuration="6.584426876s" podCreationTimestamp="2026-02-20 08:22:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:49.583555925 +0000 UTC m=+1018.558050735" watchObservedRunningTime="2026-02-20 08:22:49.584426876 +0000 UTC m=+1018.558921696" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.051082 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.060301 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152083 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-credential-keys\") pod \"d84a71b6-7503-4150-a5c4-4579b08a669a\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152146 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-config-data\") pod \"d84a71b6-7503-4150-a5c4-4579b08a669a\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152416 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-scripts\") pod \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152463 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5zcb\" (UniqueName: \"kubernetes.io/projected/e0f82f51-8a20-402d-8861-6c57a4e5af0d-kube-api-access-q5zcb\") pod \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152500 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-scripts\") pod \"d84a71b6-7503-4150-a5c4-4579b08a669a\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152522 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-httpd-run\") pod \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152546 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-combined-ca-bundle\") pod \"d84a71b6-7503-4150-a5c4-4579b08a669a\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152566 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-config-data\") pod \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152609 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-fernet-keys\") pod \"d84a71b6-7503-4150-a5c4-4579b08a669a\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152630 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152668 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-combined-ca-bundle\") pod \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152701 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jr9d\" (UniqueName: \"kubernetes.io/projected/d84a71b6-7503-4150-a5c4-4579b08a669a-kube-api-access-6jr9d\") pod \"d84a71b6-7503-4150-a5c4-4579b08a669a\" (UID: \"d84a71b6-7503-4150-a5c4-4579b08a669a\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.152718 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-logs\") pod \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\" (UID: \"e0f82f51-8a20-402d-8861-6c57a4e5af0d\") " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.154009 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-logs" (OuterVolumeSpecName: "logs") pod "e0f82f51-8a20-402d-8861-6c57a4e5af0d" (UID: "e0f82f51-8a20-402d-8861-6c57a4e5af0d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.165818 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d84a71b6-7503-4150-a5c4-4579b08a669a" (UID: "d84a71b6-7503-4150-a5c4-4579b08a669a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.166253 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "e0f82f51-8a20-402d-8861-6c57a4e5af0d" (UID: "e0f82f51-8a20-402d-8861-6c57a4e5af0d"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.173147 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0f82f51-8a20-402d-8861-6c57a4e5af0d-kube-api-access-q5zcb" (OuterVolumeSpecName: "kube-api-access-q5zcb") pod "e0f82f51-8a20-402d-8861-6c57a4e5af0d" (UID: "e0f82f51-8a20-402d-8861-6c57a4e5af0d"). InnerVolumeSpecName "kube-api-access-q5zcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.178348 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e0f82f51-8a20-402d-8861-6c57a4e5af0d" (UID: "e0f82f51-8a20-402d-8861-6c57a4e5af0d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.178347 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d84a71b6-7503-4150-a5c4-4579b08a669a" (UID: "d84a71b6-7503-4150-a5c4-4579b08a669a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.180191 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d84a71b6-7503-4150-a5c4-4579b08a669a-kube-api-access-6jr9d" (OuterVolumeSpecName: "kube-api-access-6jr9d") pod "d84a71b6-7503-4150-a5c4-4579b08a669a" (UID: "d84a71b6-7503-4150-a5c4-4579b08a669a"). InnerVolumeSpecName "kube-api-access-6jr9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.202477 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-scripts" (OuterVolumeSpecName: "scripts") pod "d84a71b6-7503-4150-a5c4-4579b08a669a" (UID: "d84a71b6-7503-4150-a5c4-4579b08a669a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.202520 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-scripts" (OuterVolumeSpecName: "scripts") pod "e0f82f51-8a20-402d-8861-6c57a4e5af0d" (UID: "e0f82f51-8a20-402d-8861-6c57a4e5af0d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.246257 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0f82f51-8a20-402d-8861-6c57a4e5af0d" (UID: "e0f82f51-8a20-402d-8861-6c57a4e5af0d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.254705 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255036 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jr9d\" (UniqueName: \"kubernetes.io/projected/d84a71b6-7503-4150-a5c4-4579b08a669a-kube-api-access-6jr9d\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255161 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255239 4948 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255315 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255370 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5zcb\" (UniqueName: \"kubernetes.io/projected/e0f82f51-8a20-402d-8861-6c57a4e5af0d-kube-api-access-q5zcb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255445 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255519 4948 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e0f82f51-8a20-402d-8861-6c57a4e5af0d-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255598 4948 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.255854 4948 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.280048 4948 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.292785 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d84a71b6-7503-4150-a5c4-4579b08a669a" (UID: "d84a71b6-7503-4150-a5c4-4579b08a669a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.306431 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-config-data" (OuterVolumeSpecName: "config-data") pod "d84a71b6-7503-4150-a5c4-4579b08a669a" (UID: "d84a71b6-7503-4150-a5c4-4579b08a669a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.322279 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-config-data" (OuterVolumeSpecName: "config-data") pod "e0f82f51-8a20-402d-8861-6c57a4e5af0d" (UID: "e0f82f51-8a20-402d-8861-6c57a4e5af0d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.358490 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.358560 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84a71b6-7503-4150-a5c4-4579b08a669a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.358573 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f82f51-8a20-402d-8861-6c57a4e5af0d-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.358583 4948 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.586404 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v29fx" event={"ID":"d84a71b6-7503-4150-a5c4-4579b08a669a","Type":"ContainerDied","Data":"21944c755e2a58a194f7e12be68919b81d0c021ee6380fb185cda3cec0fcee37"} Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.586704 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21944c755e2a58a194f7e12be68919b81d0c021ee6380fb185cda3cec0fcee37" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.586669 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v29fx" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.601254 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e0f82f51-8a20-402d-8861-6c57a4e5af0d","Type":"ContainerDied","Data":"70cd0cac387b5f659d055b8820f12fcea594407eebec35ab497b573fb1f4db5a"} Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.601308 4948 scope.go:117] "RemoveContainer" containerID="1a5b270055c8b2b5a29d54a228d2fd2e9bdac72aaebc130807e4057de5afdb1a" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.601442 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.620826 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g27hk" event={"ID":"d6729921-0a12-46a0-824f-810df88ae426","Type":"ContainerStarted","Data":"6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206"} Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.626167 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d46988dcb-rn6qs" event={"ID":"8b1ee525-1a6e-461d-aa4d-178ab601ec9d","Type":"ContainerStarted","Data":"6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af"} Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.626204 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d46988dcb-rn6qs" event={"ID":"8b1ee525-1a6e-461d-aa4d-178ab601ec9d","Type":"ContainerStarted","Data":"26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26"} Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.627134 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.627176 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.644437 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh5s2" event={"ID":"426b37f9-4acb-4d28-a55c-3fcde2600002","Type":"ContainerStarted","Data":"3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6"} Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.644490 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.665905 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.679892 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.682457 4948 scope.go:117] "RemoveContainer" containerID="7020629c426e657a58b5859f74b5d55f5c4fb964822922a0eb88d95415abe7d8" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.733712 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:50 crc kubenswrapper[4948]: E0220 08:22:50.734190 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerName="glance-log" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.734208 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerName="glance-log" Feb 20 08:22:50 crc kubenswrapper[4948]: E0220 08:22:50.734246 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerName="glance-httpd" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.734255 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerName="glance-httpd" Feb 20 08:22:50 crc kubenswrapper[4948]: E0220 08:22:50.734276 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d84a71b6-7503-4150-a5c4-4579b08a669a" containerName="keystone-bootstrap" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.734283 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d84a71b6-7503-4150-a5c4-4579b08a669a" containerName="keystone-bootstrap" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.734503 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d84a71b6-7503-4150-a5c4-4579b08a669a" containerName="keystone-bootstrap" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.734547 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerName="glance-httpd" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.734575 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" containerName="glance-log" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.735796 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.738323 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.742502 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.748537 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g27hk" podStartSLOduration=19.425732473 podStartE2EDuration="25.748512013s" podCreationTimestamp="2026-02-20 08:22:25 +0000 UTC" firstStartedPulling="2026-02-20 08:22:43.275291984 +0000 UTC m=+1012.249786794" lastFinishedPulling="2026-02-20 08:22:49.598071514 +0000 UTC m=+1018.572566334" observedRunningTime="2026-02-20 08:22:50.660462804 +0000 UTC m=+1019.634957634" watchObservedRunningTime="2026-02-20 08:22:50.748512013 +0000 UTC m=+1019.723006823" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.763173 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.766915 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-d46988dcb-rn6qs" podStartSLOduration=3.766895298 podStartE2EDuration="3.766895298s" podCreationTimestamp="2026-02-20 08:22:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:50.688212221 +0000 UTC m=+1019.662707041" watchObservedRunningTime="2026-02-20 08:22:50.766895298 +0000 UTC m=+1019.741390118" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.779030 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6d4b8df464-wspg2"] Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.780153 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.780631 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hh5s2" podStartSLOduration=16.580718558 podStartE2EDuration="22.780611937s" podCreationTimestamp="2026-02-20 08:22:28 +0000 UTC" firstStartedPulling="2026-02-20 08:22:43.275410777 +0000 UTC m=+1012.249905597" lastFinishedPulling="2026-02-20 08:22:49.475304156 +0000 UTC m=+1018.449798976" observedRunningTime="2026-02-20 08:22:50.719482804 +0000 UTC m=+1019.693977625" watchObservedRunningTime="2026-02-20 08:22:50.780611937 +0000 UTC m=+1019.755106757" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.788506 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-v9rj2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.788689 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.789016 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.789154 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.789308 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.789434 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.861017 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6d4b8df464-wspg2"] Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874059 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874150 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-config-data\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874246 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-credential-keys\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874276 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874301 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-fernet-keys\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874332 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-combined-ca-bundle\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874366 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-public-tls-certs\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874401 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874455 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-scripts\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874478 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-logs\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874510 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44bpr\" (UniqueName: \"kubernetes.io/projected/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-kube-api-access-44bpr\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874541 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdk7s\" (UniqueName: \"kubernetes.io/projected/d9399b9c-0935-4735-8b08-96d4a29d4ba8-kube-api-access-jdk7s\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874631 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874653 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-scripts\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874676 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-config-data\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.874696 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-internal-tls-certs\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.998912 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.998999 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-scripts\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999045 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-config-data\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999086 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-internal-tls-certs\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999168 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999260 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-config-data\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999407 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-credential-keys\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999452 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999490 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-fernet-keys\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999542 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-combined-ca-bundle\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999606 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-public-tls-certs\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999669 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999731 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-scripts\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999760 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-logs\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999817 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44bpr\" (UniqueName: \"kubernetes.io/projected/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-kube-api-access-44bpr\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:50 crc kubenswrapper[4948]: I0220 08:22:50.999853 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdk7s\" (UniqueName: \"kubernetes.io/projected/d9399b9c-0935-4735-8b08-96d4a29d4ba8-kube-api-access-jdk7s\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.006035 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-scripts\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.008188 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-internal-tls-certs\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.008718 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-config-data\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.011606 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.013731 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.015360 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-public-tls-certs\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.020527 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-credential-keys\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.027281 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-logs\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.032144 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-combined-ca-bundle\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.035622 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.035633 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-fernet-keys\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.038500 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.039570 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-config-data\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.041985 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44bpr\" (UniqueName: \"kubernetes.io/projected/6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc-kube-api-access-44bpr\") pod \"keystone-6d4b8df464-wspg2\" (UID: \"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc\") " pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.045628 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdk7s\" (UniqueName: \"kubernetes.io/projected/d9399b9c-0935-4735-8b08-96d4a29d4ba8-kube-api-access-jdk7s\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.046610 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-scripts\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.077840 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.116525 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.363452 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.691552 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6d4b8df464-wspg2"] Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.760386 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0f82f51-8a20-402d-8861-6c57a4e5af0d" path="/var/lib/kubelet/pods/e0f82f51-8a20-402d-8861-6c57a4e5af0d/volumes" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.812250 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.876364 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-96668"] Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.876589 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" podUID="94b82d55-b467-408f-8eb1-6da4648dde97" containerName="dnsmasq-dns" containerID="cri-o://76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0" gracePeriod=10 Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.980044 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:51 crc kubenswrapper[4948]: I0220 08:22:51.980100 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.019854 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.038202 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.136427 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.396304 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.542539 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-nb\") pod \"94b82d55-b467-408f-8eb1-6da4648dde97\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.542616 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-sb\") pod \"94b82d55-b467-408f-8eb1-6da4648dde97\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.542771 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-swift-storage-0\") pod \"94b82d55-b467-408f-8eb1-6da4648dde97\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.542810 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-svc\") pod \"94b82d55-b467-408f-8eb1-6da4648dde97\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.542864 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-config\") pod \"94b82d55-b467-408f-8eb1-6da4648dde97\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.542887 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqddr\" (UniqueName: \"kubernetes.io/projected/94b82d55-b467-408f-8eb1-6da4648dde97-kube-api-access-rqddr\") pod \"94b82d55-b467-408f-8eb1-6da4648dde97\" (UID: \"94b82d55-b467-408f-8eb1-6da4648dde97\") " Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.554939 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94b82d55-b467-408f-8eb1-6da4648dde97-kube-api-access-rqddr" (OuterVolumeSpecName: "kube-api-access-rqddr") pod "94b82d55-b467-408f-8eb1-6da4648dde97" (UID: "94b82d55-b467-408f-8eb1-6da4648dde97"). InnerVolumeSpecName "kube-api-access-rqddr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.591136 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "94b82d55-b467-408f-8eb1-6da4648dde97" (UID: "94b82d55-b467-408f-8eb1-6da4648dde97"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.612888 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "94b82d55-b467-408f-8eb1-6da4648dde97" (UID: "94b82d55-b467-408f-8eb1-6da4648dde97"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.616413 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "94b82d55-b467-408f-8eb1-6da4648dde97" (UID: "94b82d55-b467-408f-8eb1-6da4648dde97"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.616531 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "94b82d55-b467-408f-8eb1-6da4648dde97" (UID: "94b82d55-b467-408f-8eb1-6da4648dde97"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.645334 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqddr\" (UniqueName: \"kubernetes.io/projected/94b82d55-b467-408f-8eb1-6da4648dde97-kube-api-access-rqddr\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.645415 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.645429 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.645442 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.645515 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.654497 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-config" (OuterVolumeSpecName: "config") pod "94b82d55-b467-408f-8eb1-6da4648dde97" (UID: "94b82d55-b467-408f-8eb1-6da4648dde97"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.708491 4948 generic.go:334] "Generic (PLEG): container finished" podID="94b82d55-b467-408f-8eb1-6da4648dde97" containerID="76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0" exitCode=0 Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.708788 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" event={"ID":"94b82d55-b467-408f-8eb1-6da4648dde97","Type":"ContainerDied","Data":"76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0"} Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.708822 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" event={"ID":"94b82d55-b467-408f-8eb1-6da4648dde97","Type":"ContainerDied","Data":"af3aca3908294e0c1fe45deb6418a29e5d516dc4d54d54ff11acdb6845abd301"} Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.708845 4948 scope.go:117] "RemoveContainer" containerID="76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.709014 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-96668" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.713181 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d9399b9c-0935-4735-8b08-96d4a29d4ba8","Type":"ContainerStarted","Data":"fecc6864756af4b1a18a9eaf7ed69fde086ccbc2685b82cc1f598b9178cb31a0"} Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.724409 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6d4b8df464-wspg2" event={"ID":"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc","Type":"ContainerStarted","Data":"6c85d3af2063b59d7ee0e44b6d6917212c7171f281d6864ee9d371d95e5f223a"} Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.733376 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6d4b8df464-wspg2" event={"ID":"6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc","Type":"ContainerStarted","Data":"ad469bae9f6d71112f325de3274ec1b7fb44241d69e1e8658079429caacb8930"} Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.733666 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.734161 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.746796 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94b82d55-b467-408f-8eb1-6da4648dde97-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.750991 4948 scope.go:117] "RemoveContainer" containerID="bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.765901 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6d4b8df464-wspg2" podStartSLOduration=2.765886707 podStartE2EDuration="2.765886707s" podCreationTimestamp="2026-02-20 08:22:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:52.749384369 +0000 UTC m=+1021.723879209" watchObservedRunningTime="2026-02-20 08:22:52.765886707 +0000 UTC m=+1021.740381527" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.783039 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-96668"] Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.793139 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-96668"] Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.814387 4948 scope.go:117] "RemoveContainer" containerID="76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0" Feb 20 08:22:52 crc kubenswrapper[4948]: E0220 08:22:52.814826 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0\": container with ID starting with 76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0 not found: ID does not exist" containerID="76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.814856 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0"} err="failed to get container status \"76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0\": rpc error: code = NotFound desc = could not find container \"76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0\": container with ID starting with 76e608849a7aba9c1d66b8ca878674048ebd2cf05258476f0e82590946ed89c0 not found: ID does not exist" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.814875 4948 scope.go:117] "RemoveContainer" containerID="bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df" Feb 20 08:22:52 crc kubenswrapper[4948]: E0220 08:22:52.815193 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df\": container with ID starting with bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df not found: ID does not exist" containerID="bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df" Feb 20 08:22:52 crc kubenswrapper[4948]: I0220 08:22:52.815212 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df"} err="failed to get container status \"bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df\": rpc error: code = NotFound desc = could not find container \"bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df\": container with ID starting with bbaf3121e36eabf03661c450060f71b1a26f424ea932363040deddf76d19e0df not found: ID does not exist" Feb 20 08:22:53 crc kubenswrapper[4948]: I0220 08:22:53.736054 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94b82d55-b467-408f-8eb1-6da4648dde97" path="/var/lib/kubelet/pods/94b82d55-b467-408f-8eb1-6da4648dde97/volumes" Feb 20 08:22:53 crc kubenswrapper[4948]: I0220 08:22:53.739271 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d9399b9c-0935-4735-8b08-96d4a29d4ba8","Type":"ContainerStarted","Data":"656e84496ca041e8a454b07ed67e165b130af1030016564ae4519b8e2e6f11e5"} Feb 20 08:22:53 crc kubenswrapper[4948]: I0220 08:22:53.739559 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d9399b9c-0935-4735-8b08-96d4a29d4ba8","Type":"ContainerStarted","Data":"a66ddb00a7f57dc492a3a8a03a3bce06f45c65381b1e5db4f923e10fec5267a0"} Feb 20 08:22:53 crc kubenswrapper[4948]: I0220 08:22:53.740272 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:22:53 crc kubenswrapper[4948]: I0220 08:22:53.768647 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.768628852 podStartE2EDuration="3.768628852s" podCreationTimestamp="2026-02-20 08:22:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:22:53.755904787 +0000 UTC m=+1022.730399617" watchObservedRunningTime="2026-02-20 08:22:53.768628852 +0000 UTC m=+1022.743123672" Feb 20 08:22:54 crc kubenswrapper[4948]: I0220 08:22:54.747012 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:22:54 crc kubenswrapper[4948]: I0220 08:22:54.747040 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:22:54 crc kubenswrapper[4948]: I0220 08:22:54.779914 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:54 crc kubenswrapper[4948]: I0220 08:22:54.781383 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 20 08:22:55 crc kubenswrapper[4948]: I0220 08:22:55.478886 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-66b4d4cb8-hgmkd" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Feb 20 08:22:55 crc kubenswrapper[4948]: I0220 08:22:55.646403 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-685c4db87b-rw57h" podUID="7b9a072a-a040-414c-96c7-cf1454148745" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Feb 20 08:22:55 crc kubenswrapper[4948]: I0220 08:22:55.966929 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:55 crc kubenswrapper[4948]: I0220 08:22:55.967466 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:56 crc kubenswrapper[4948]: I0220 08:22:56.013505 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:56 crc kubenswrapper[4948]: I0220 08:22:56.845068 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:57 crc kubenswrapper[4948]: I0220 08:22:57.249714 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g27hk"] Feb 20 08:22:57 crc kubenswrapper[4948]: I0220 08:22:57.814588 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-c7zzv" event={"ID":"f1f108e1-27bb-433e-b7e1-f9cc7f778182","Type":"ContainerStarted","Data":"b6240fac3873ea6c61937ec67473ce610d9e47448d45e4e229e635b6cb80cfa1"} Feb 20 08:22:57 crc kubenswrapper[4948]: I0220 08:22:57.820398 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerStarted","Data":"61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07"} Feb 20 08:22:57 crc kubenswrapper[4948]: I0220 08:22:57.843832 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-c7zzv" podStartSLOduration=3.872458231 podStartE2EDuration="43.843800229s" podCreationTimestamp="2026-02-20 08:22:14 +0000 UTC" firstStartedPulling="2026-02-20 08:22:16.603201852 +0000 UTC m=+985.577696682" lastFinishedPulling="2026-02-20 08:22:56.57454386 +0000 UTC m=+1025.549038680" observedRunningTime="2026-02-20 08:22:57.829053574 +0000 UTC m=+1026.803548414" watchObservedRunningTime="2026-02-20 08:22:57.843800229 +0000 UTC m=+1026.818295079" Feb 20 08:22:58 crc kubenswrapper[4948]: I0220 08:22:58.829192 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g27hk" podUID="d6729921-0a12-46a0-824f-810df88ae426" containerName="registry-server" containerID="cri-o://6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206" gracePeriod=2 Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.307666 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.358651 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.358701 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.370184 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2k5c\" (UniqueName: \"kubernetes.io/projected/d6729921-0a12-46a0-824f-810df88ae426-kube-api-access-x2k5c\") pod \"d6729921-0a12-46a0-824f-810df88ae426\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.370291 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-utilities\") pod \"d6729921-0a12-46a0-824f-810df88ae426\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.370355 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-catalog-content\") pod \"d6729921-0a12-46a0-824f-810df88ae426\" (UID: \"d6729921-0a12-46a0-824f-810df88ae426\") " Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.371414 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-utilities" (OuterVolumeSpecName: "utilities") pod "d6729921-0a12-46a0-824f-810df88ae426" (UID: "d6729921-0a12-46a0-824f-810df88ae426"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.376014 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6729921-0a12-46a0-824f-810df88ae426-kube-api-access-x2k5c" (OuterVolumeSpecName: "kube-api-access-x2k5c") pod "d6729921-0a12-46a0-824f-810df88ae426" (UID: "d6729921-0a12-46a0-824f-810df88ae426"). InnerVolumeSpecName "kube-api-access-x2k5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.428320 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6729921-0a12-46a0-824f-810df88ae426" (UID: "d6729921-0a12-46a0-824f-810df88ae426"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.472861 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2k5c\" (UniqueName: \"kubernetes.io/projected/d6729921-0a12-46a0-824f-810df88ae426-kube-api-access-x2k5c\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.472897 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.472906 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6729921-0a12-46a0-824f-810df88ae426-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.847407 4948 generic.go:334] "Generic (PLEG): container finished" podID="d6729921-0a12-46a0-824f-810df88ae426" containerID="6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206" exitCode=0 Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.847481 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g27hk" event={"ID":"d6729921-0a12-46a0-824f-810df88ae426","Type":"ContainerDied","Data":"6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206"} Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.847512 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g27hk" event={"ID":"d6729921-0a12-46a0-824f-810df88ae426","Type":"ContainerDied","Data":"b7570c60dc7ea9b2dda7deeb7d6ef87e7350a05f972449d37dfa272d75ce2941"} Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.847532 4948 scope.go:117] "RemoveContainer" containerID="6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.847679 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g27hk" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.854662 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xt2zq" event={"ID":"f706b741-86f9-49a3-95d5-85ee62eb3668","Type":"ContainerStarted","Data":"5b0b3d5fbb40595687dd65feb8573194de8a92b25615987166c82509c29136f4"} Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.857457 4948 generic.go:334] "Generic (PLEG): container finished" podID="f1f108e1-27bb-433e-b7e1-f9cc7f778182" containerID="b6240fac3873ea6c61937ec67473ce610d9e47448d45e4e229e635b6cb80cfa1" exitCode=0 Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.857501 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-c7zzv" event={"ID":"f1f108e1-27bb-433e-b7e1-f9cc7f778182","Type":"ContainerDied","Data":"b6240fac3873ea6c61937ec67473ce610d9e47448d45e4e229e635b6cb80cfa1"} Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.887172 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g27hk"] Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.888955 4948 scope.go:117] "RemoveContainer" containerID="bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.905123 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g27hk"] Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.914400 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-xt2zq" podStartSLOduration=3.030407265 podStartE2EDuration="45.91437863s" podCreationTimestamp="2026-02-20 08:22:14 +0000 UTC" firstStartedPulling="2026-02-20 08:22:15.460074934 +0000 UTC m=+984.434569754" lastFinishedPulling="2026-02-20 08:22:58.344046299 +0000 UTC m=+1027.318541119" observedRunningTime="2026-02-20 08:22:59.88648426 +0000 UTC m=+1028.860979080" watchObservedRunningTime="2026-02-20 08:22:59.91437863 +0000 UTC m=+1028.888873460" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.925592 4948 scope.go:117] "RemoveContainer" containerID="c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.955082 4948 scope.go:117] "RemoveContainer" containerID="6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206" Feb 20 08:22:59 crc kubenswrapper[4948]: E0220 08:22:59.956067 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206\": container with ID starting with 6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206 not found: ID does not exist" containerID="6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.956104 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206"} err="failed to get container status \"6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206\": rpc error: code = NotFound desc = could not find container \"6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206\": container with ID starting with 6cd76a3b8281c950d8cac3fc54816ec3fe45fb413dc79249c79a1f2aecbbd206 not found: ID does not exist" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.956128 4948 scope.go:117] "RemoveContainer" containerID="bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793" Feb 20 08:22:59 crc kubenswrapper[4948]: E0220 08:22:59.957183 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793\": container with ID starting with bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793 not found: ID does not exist" containerID="bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.957203 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793"} err="failed to get container status \"bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793\": rpc error: code = NotFound desc = could not find container \"bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793\": container with ID starting with bdb7ba6b04df20596c79e2a31fb3e44aca83f1668de6adef1942e6b1ba6fa793 not found: ID does not exist" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.957222 4948 scope.go:117] "RemoveContainer" containerID="c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a" Feb 20 08:22:59 crc kubenswrapper[4948]: E0220 08:22:59.957479 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a\": container with ID starting with c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a not found: ID does not exist" containerID="c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a" Feb 20 08:22:59 crc kubenswrapper[4948]: I0220 08:22:59.957566 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a"} err="failed to get container status \"c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a\": rpc error: code = NotFound desc = could not find container \"c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a\": container with ID starting with c30e42a1011ef0a11a3bd3a8a1babce0958085ece91112de32b9f174e1d1c87a not found: ID does not exist" Feb 20 08:23:00 crc kubenswrapper[4948]: I0220 08:23:00.407305 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-hh5s2" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="registry-server" probeResult="failure" output=< Feb 20 08:23:00 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 08:23:00 crc kubenswrapper[4948]: > Feb 20 08:23:01 crc kubenswrapper[4948]: I0220 08:23:01.364694 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 20 08:23:01 crc kubenswrapper[4948]: I0220 08:23:01.364836 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 20 08:23:01 crc kubenswrapper[4948]: I0220 08:23:01.406322 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 20 08:23:01 crc kubenswrapper[4948]: I0220 08:23:01.419667 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 20 08:23:01 crc kubenswrapper[4948]: I0220 08:23:01.737625 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6729921-0a12-46a0-824f-810df88ae426" path="/var/lib/kubelet/pods/d6729921-0a12-46a0-824f-810df88ae426/volumes" Feb 20 08:23:01 crc kubenswrapper[4948]: I0220 08:23:01.883442 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 20 08:23:01 crc kubenswrapper[4948]: I0220 08:23:01.883498 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 20 08:23:02 crc kubenswrapper[4948]: I0220 08:23:02.889174 4948 generic.go:334] "Generic (PLEG): container finished" podID="f706b741-86f9-49a3-95d5-85ee62eb3668" containerID="5b0b3d5fbb40595687dd65feb8573194de8a92b25615987166c82509c29136f4" exitCode=0 Feb 20 08:23:02 crc kubenswrapper[4948]: I0220 08:23:02.889957 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xt2zq" event={"ID":"f706b741-86f9-49a3-95d5-85ee62eb3668","Type":"ContainerDied","Data":"5b0b3d5fbb40595687dd65feb8573194de8a92b25615987166c82509c29136f4"} Feb 20 08:23:03 crc kubenswrapper[4948]: I0220 08:23:03.641287 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 20 08:23:03 crc kubenswrapper[4948]: I0220 08:23:03.757400 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.883523 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.950216 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-c7zzv" event={"ID":"f1f108e1-27bb-433e-b7e1-f9cc7f778182","Type":"ContainerDied","Data":"1c7c890826add7e27d4d21f7cb357370f931db1e77090c60ac1c40f2ed6d57cc"} Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.950265 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c7c890826add7e27d4d21f7cb357370f931db1e77090c60ac1c40f2ed6d57cc" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.972164 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xt2zq" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.972331 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xt2zq" event={"ID":"f706b741-86f9-49a3-95d5-85ee62eb3668","Type":"ContainerDied","Data":"571bfb34e93031079bb93437b09bfef171c7529235d603237690b075a9933988"} Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.972352 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="571bfb34e93031079bb93437b09bfef171c7529235d603237690b075a9933988" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.983334 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f706b741-86f9-49a3-95d5-85ee62eb3668-etc-machine-id\") pod \"f706b741-86f9-49a3-95d5-85ee62eb3668\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.983521 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-scripts\") pod \"f706b741-86f9-49a3-95d5-85ee62eb3668\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.983453 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f706b741-86f9-49a3-95d5-85ee62eb3668-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f706b741-86f9-49a3-95d5-85ee62eb3668" (UID: "f706b741-86f9-49a3-95d5-85ee62eb3668"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.983664 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-config-data\") pod \"f706b741-86f9-49a3-95d5-85ee62eb3668\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.983777 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwtv5\" (UniqueName: \"kubernetes.io/projected/f706b741-86f9-49a3-95d5-85ee62eb3668-kube-api-access-fwtv5\") pod \"f706b741-86f9-49a3-95d5-85ee62eb3668\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.983888 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-db-sync-config-data\") pod \"f706b741-86f9-49a3-95d5-85ee62eb3668\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.983981 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-combined-ca-bundle\") pod \"f706b741-86f9-49a3-95d5-85ee62eb3668\" (UID: \"f706b741-86f9-49a3-95d5-85ee62eb3668\") " Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.985614 4948 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f706b741-86f9-49a3-95d5-85ee62eb3668-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.987466 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-scripts" (OuterVolumeSpecName: "scripts") pod "f706b741-86f9-49a3-95d5-85ee62eb3668" (UID: "f706b741-86f9-49a3-95d5-85ee62eb3668"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.988164 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f706b741-86f9-49a3-95d5-85ee62eb3668-kube-api-access-fwtv5" (OuterVolumeSpecName: "kube-api-access-fwtv5") pod "f706b741-86f9-49a3-95d5-85ee62eb3668" (UID: "f706b741-86f9-49a3-95d5-85ee62eb3668"). InnerVolumeSpecName "kube-api-access-fwtv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:04 crc kubenswrapper[4948]: I0220 08:23:04.988552 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f706b741-86f9-49a3-95d5-85ee62eb3668" (UID: "f706b741-86f9-49a3-95d5-85ee62eb3668"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.008866 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f706b741-86f9-49a3-95d5-85ee62eb3668" (UID: "f706b741-86f9-49a3-95d5-85ee62eb3668"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.024776 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-config-data" (OuterVolumeSpecName: "config-data") pod "f706b741-86f9-49a3-95d5-85ee62eb3668" (UID: "f706b741-86f9-49a3-95d5-85ee62eb3668"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.042633 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.086549 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-combined-ca-bundle\") pod \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.086676 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-db-sync-config-data\") pod \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.086777 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmlgz\" (UniqueName: \"kubernetes.io/projected/f1f108e1-27bb-433e-b7e1-f9cc7f778182-kube-api-access-gmlgz\") pod \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\" (UID: \"f1f108e1-27bb-433e-b7e1-f9cc7f778182\") " Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.087337 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.087360 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.087867 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwtv5\" (UniqueName: \"kubernetes.io/projected/f706b741-86f9-49a3-95d5-85ee62eb3668-kube-api-access-fwtv5\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.087884 4948 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.087896 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f706b741-86f9-49a3-95d5-85ee62eb3668-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.092034 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f1f108e1-27bb-433e-b7e1-f9cc7f778182" (UID: "f1f108e1-27bb-433e-b7e1-f9cc7f778182"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.093743 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1f108e1-27bb-433e-b7e1-f9cc7f778182-kube-api-access-gmlgz" (OuterVolumeSpecName: "kube-api-access-gmlgz") pod "f1f108e1-27bb-433e-b7e1-f9cc7f778182" (UID: "f1f108e1-27bb-433e-b7e1-f9cc7f778182"). InnerVolumeSpecName "kube-api-access-gmlgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.138901 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1f108e1-27bb-433e-b7e1-f9cc7f778182" (UID: "f1f108e1-27bb-433e-b7e1-f9cc7f778182"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.189324 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.189359 4948 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f1f108e1-27bb-433e-b7e1-f9cc7f778182-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.189372 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmlgz\" (UniqueName: \"kubernetes.io/projected/f1f108e1-27bb-433e-b7e1-f9cc7f778182-kube-api-access-gmlgz\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.221098 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:05 crc kubenswrapper[4948]: E0220 08:23:05.221699 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1f108e1-27bb-433e-b7e1-f9cc7f778182" containerName="barbican-db-sync" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.221804 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1f108e1-27bb-433e-b7e1-f9cc7f778182" containerName="barbican-db-sync" Feb 20 08:23:05 crc kubenswrapper[4948]: E0220 08:23:05.221867 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94b82d55-b467-408f-8eb1-6da4648dde97" containerName="init" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.221918 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="94b82d55-b467-408f-8eb1-6da4648dde97" containerName="init" Feb 20 08:23:05 crc kubenswrapper[4948]: E0220 08:23:05.221996 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94b82d55-b467-408f-8eb1-6da4648dde97" containerName="dnsmasq-dns" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222059 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="94b82d55-b467-408f-8eb1-6da4648dde97" containerName="dnsmasq-dns" Feb 20 08:23:05 crc kubenswrapper[4948]: E0220 08:23:05.222127 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6729921-0a12-46a0-824f-810df88ae426" containerName="extract-content" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222181 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6729921-0a12-46a0-824f-810df88ae426" containerName="extract-content" Feb 20 08:23:05 crc kubenswrapper[4948]: E0220 08:23:05.222242 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6729921-0a12-46a0-824f-810df88ae426" containerName="extract-utilities" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222290 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6729921-0a12-46a0-824f-810df88ae426" containerName="extract-utilities" Feb 20 08:23:05 crc kubenswrapper[4948]: E0220 08:23:05.222345 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f706b741-86f9-49a3-95d5-85ee62eb3668" containerName="cinder-db-sync" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222413 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f706b741-86f9-49a3-95d5-85ee62eb3668" containerName="cinder-db-sync" Feb 20 08:23:05 crc kubenswrapper[4948]: E0220 08:23:05.222482 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6729921-0a12-46a0-824f-810df88ae426" containerName="registry-server" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222534 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6729921-0a12-46a0-824f-810df88ae426" containerName="registry-server" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222732 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="94b82d55-b467-408f-8eb1-6da4648dde97" containerName="dnsmasq-dns" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222790 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6729921-0a12-46a0-824f-810df88ae426" containerName="registry-server" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222843 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1f108e1-27bb-433e-b7e1-f9cc7f778182" containerName="barbican-db-sync" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.222910 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f706b741-86f9-49a3-95d5-85ee62eb3668" containerName="cinder-db-sync" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.223808 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.229362 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.243121 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.267016 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b895b5785-97gsc"] Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.268481 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.337507 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-97gsc"] Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.391963 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392043 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-config\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392067 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-svc\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392092 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392115 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vllr\" (UniqueName: \"kubernetes.io/projected/75294aee-c7f3-4b0b-82fc-df058490b90f-kube-api-access-8vllr\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392156 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392176 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392230 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392255 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6wdv\" (UniqueName: \"kubernetes.io/projected/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-kube-api-access-c6wdv\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392327 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392365 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75294aee-c7f3-4b0b-82fc-df058490b90f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.392386 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-scripts\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.478060 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-66b4d4cb8-hgmkd" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494239 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494290 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494328 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494355 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6wdv\" (UniqueName: \"kubernetes.io/projected/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-kube-api-access-c6wdv\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494382 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494407 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75294aee-c7f3-4b0b-82fc-df058490b90f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494422 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-scripts\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494467 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494499 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-config\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494523 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-svc\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494548 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494567 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vllr\" (UniqueName: \"kubernetes.io/projected/75294aee-c7f3-4b0b-82fc-df058490b90f-kube-api-access-8vllr\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.494836 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75294aee-c7f3-4b0b-82fc-df058490b90f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.496141 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-svc\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.496192 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-config\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.496282 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.497796 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.499242 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.500480 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-scripts\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.502495 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.504805 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.508471 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.517606 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vllr\" (UniqueName: \"kubernetes.io/projected/75294aee-c7f3-4b0b-82fc-df058490b90f-kube-api-access-8vllr\") pod \"cinder-scheduler-0\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.521893 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6wdv\" (UniqueName: \"kubernetes.io/projected/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-kube-api-access-c6wdv\") pod \"dnsmasq-dns-b895b5785-97gsc\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.561827 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.563634 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.565407 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.571082 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.573045 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.623351 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.715820 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq6j4\" (UniqueName: \"kubernetes.io/projected/1ff6993b-2bde-429d-b0c7-cdea51f733a2-kube-api-access-xq6j4\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.715888 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data-custom\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.715949 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.715984 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.716005 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ff6993b-2bde-429d-b0c7-cdea51f733a2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.716037 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ff6993b-2bde-429d-b0c7-cdea51f733a2-logs\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.716062 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-scripts\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.817789 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq6j4\" (UniqueName: \"kubernetes.io/projected/1ff6993b-2bde-429d-b0c7-cdea51f733a2-kube-api-access-xq6j4\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.817865 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data-custom\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.817941 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.817962 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.817994 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ff6993b-2bde-429d-b0c7-cdea51f733a2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.818029 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ff6993b-2bde-429d-b0c7-cdea51f733a2-logs\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.818046 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-scripts\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.820213 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ff6993b-2bde-429d-b0c7-cdea51f733a2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.821909 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ff6993b-2bde-429d-b0c7-cdea51f733a2-logs\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.825073 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-scripts\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.831901 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data-custom\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.833162 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.834674 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq6j4\" (UniqueName: \"kubernetes.io/projected/1ff6993b-2bde-429d-b0c7-cdea51f733a2-kube-api-access-xq6j4\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.842160 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.887683 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 08:23:05 crc kubenswrapper[4948]: I0220 08:23:05.948473 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-97gsc"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.000205 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-97gsc" event={"ID":"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e","Type":"ContainerStarted","Data":"a73968f29763e85cb2562473f92370ee9d88e9ecbb9c1dd9bc348a65b2728c01"} Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.072435 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-c7zzv" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.072519 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerStarted","Data":"27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9"} Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.080345 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="ceilometer-central-agent" containerID="cri-o://f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56" gracePeriod=30 Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.080661 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.080961 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="proxy-httpd" containerID="cri-o://27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9" gracePeriod=30 Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.081041 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="sg-core" containerID="cri-o://61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07" gracePeriod=30 Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.081104 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="ceilometer-notification-agent" containerID="cri-o://68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493" gracePeriod=30 Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.089331 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:06 crc kubenswrapper[4948]: W0220 08:23:06.125081 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75294aee_c7f3_4b0b_82fc_df058490b90f.slice/crio-3927e56027e8aee8bf743a6fd812a761cbedf51a0e339d17550adee912677f5f WatchSource:0}: Error finding container 3927e56027e8aee8bf743a6fd812a761cbedf51a0e339d17550adee912677f5f: Status 404 returned error can't find the container with id 3927e56027e8aee8bf743a6fd812a761cbedf51a0e339d17550adee912677f5f Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.142962 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.00916312 podStartE2EDuration="52.142938327s" podCreationTimestamp="2026-02-20 08:22:14 +0000 UTC" firstStartedPulling="2026-02-20 08:22:15.726795425 +0000 UTC m=+984.701290245" lastFinishedPulling="2026-02-20 08:23:04.860570622 +0000 UTC m=+1033.835065452" observedRunningTime="2026-02-20 08:23:06.106921766 +0000 UTC m=+1035.081416596" watchObservedRunningTime="2026-02-20 08:23:06.142938327 +0000 UTC m=+1035.117433147" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.249026 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-76d486c65f-w9lc6"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.251097 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.253826 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8rh2b" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.254467 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.254779 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.266266 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6ffd468-xgk27"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.268272 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.271296 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.281233 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-76d486c65f-w9lc6"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.317787 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6ffd468-xgk27"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.340913 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-combined-ca-bundle\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.340962 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-config-data-custom\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.341031 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-config-data-custom\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.341060 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-config-data\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.341092 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/292d8e1e-de3b-4259-afef-c19e2c21187c-logs\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.341117 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz4wb\" (UniqueName: \"kubernetes.io/projected/292d8e1e-de3b-4259-afef-c19e2c21187c-kube-api-access-vz4wb\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.341140 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-combined-ca-bundle\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.341160 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b356058-d957-4fa5-a106-3fe5052c8539-logs\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.341183 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65fwq\" (UniqueName: \"kubernetes.io/projected/4b356058-d957-4fa5-a106-3fe5052c8539-kube-api-access-65fwq\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.341206 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-config-data\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.346121 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-97gsc"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.361286 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5dgw7"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.368446 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.382180 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5dgw7"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.421042 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446518 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b356058-d957-4fa5-a106-3fe5052c8539-logs\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446569 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65fwq\" (UniqueName: \"kubernetes.io/projected/4b356058-d957-4fa5-a106-3fe5052c8539-kube-api-access-65fwq\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446596 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-config-data\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446627 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446649 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-config\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446696 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-combined-ca-bundle\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446721 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-config-data-custom\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446745 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2cgj\" (UniqueName: \"kubernetes.io/projected/63124eef-9456-44d1-a8ba-cc8297af3e9b-kube-api-access-l2cgj\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446770 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446795 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-config-data-custom\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446829 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-config-data\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446852 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446874 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/292d8e1e-de3b-4259-afef-c19e2c21187c-logs\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446900 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446929 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz4wb\" (UniqueName: \"kubernetes.io/projected/292d8e1e-de3b-4259-afef-c19e2c21187c-kube-api-access-vz4wb\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.446961 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-combined-ca-bundle\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.448139 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b356058-d957-4fa5-a106-3fe5052c8539-logs\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.453480 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/292d8e1e-de3b-4259-afef-c19e2c21187c-logs\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.459184 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-combined-ca-bundle\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.459183 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-config-data-custom\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.461370 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-combined-ca-bundle\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.461891 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-config-data-custom\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.462859 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b356058-d957-4fa5-a106-3fe5052c8539-config-data\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.468626 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65fwq\" (UniqueName: \"kubernetes.io/projected/4b356058-d957-4fa5-a106-3fe5052c8539-kube-api-access-65fwq\") pod \"barbican-worker-76d486c65f-w9lc6\" (UID: \"4b356058-d957-4fa5-a106-3fe5052c8539\") " pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.468984 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/292d8e1e-de3b-4259-afef-c19e2c21187c-config-data\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.475288 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz4wb\" (UniqueName: \"kubernetes.io/projected/292d8e1e-de3b-4259-afef-c19e2c21187c-kube-api-access-vz4wb\") pod \"barbican-keystone-listener-6ffd468-xgk27\" (UID: \"292d8e1e-de3b-4259-afef-c19e2c21187c\") " pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.499326 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-65487cc4d6-7gdvr"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.503673 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.507396 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.520613 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65487cc4d6-7gdvr"] Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.549375 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.549423 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-config\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.549486 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2cgj\" (UniqueName: \"kubernetes.io/projected/63124eef-9456-44d1-a8ba-cc8297af3e9b-kube-api-access-l2cgj\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.549520 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.549582 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.549643 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.551742 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.552287 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.552868 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-config\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.554268 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.555374 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.571403 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2cgj\" (UniqueName: \"kubernetes.io/projected/63124eef-9456-44d1-a8ba-cc8297af3e9b-kube-api-access-l2cgj\") pod \"dnsmasq-dns-5c9776ccc5-5dgw7\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.597895 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76d486c65f-w9lc6" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.613670 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6ffd468-xgk27" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.650749 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld59l\" (UniqueName: \"kubernetes.io/projected/6a669dd0-d287-4d07-9eec-aea5ed80c791-kube-api-access-ld59l\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.650801 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data-custom\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.650822 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-combined-ca-bundle\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.651248 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a669dd0-d287-4d07-9eec-aea5ed80c791-logs\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.651372 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.691826 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.760992 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a669dd0-d287-4d07-9eec-aea5ed80c791-logs\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.761044 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.761075 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld59l\" (UniqueName: \"kubernetes.io/projected/6a669dd0-d287-4d07-9eec-aea5ed80c791-kube-api-access-ld59l\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.761108 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data-custom\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.761127 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-combined-ca-bundle\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.762350 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a669dd0-d287-4d07-9eec-aea5ed80c791-logs\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.767192 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data-custom\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.767391 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-combined-ca-bundle\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.768570 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.785232 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld59l\" (UniqueName: \"kubernetes.io/projected/6a669dd0-d287-4d07-9eec-aea5ed80c791-kube-api-access-ld59l\") pod \"barbican-api-65487cc4d6-7gdvr\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:06 crc kubenswrapper[4948]: I0220 08:23:06.828051 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.059503 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-76d486c65f-w9lc6"] Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.101983 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76d486c65f-w9lc6" event={"ID":"4b356058-d957-4fa5-a106-3fe5052c8539","Type":"ContainerStarted","Data":"6d267613f41bda4b9e00a369fc1b5681c20f7663cecb7ffde0195009376088c3"} Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.109793 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1ff6993b-2bde-429d-b0c7-cdea51f733a2","Type":"ContainerStarted","Data":"b0f8def01191a21afa9e066696a67301fbaaf496e33dd51ae806612c580bc882"} Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.112887 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"75294aee-c7f3-4b0b-82fc-df058490b90f","Type":"ContainerStarted","Data":"3927e56027e8aee8bf743a6fd812a761cbedf51a0e339d17550adee912677f5f"} Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.121808 4948 generic.go:334] "Generic (PLEG): container finished" podID="4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" containerID="a054eac97e88cce1aa9882153236f28817412a343a3608ef402b154e16e8b759" exitCode=0 Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.121927 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-97gsc" event={"ID":"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e","Type":"ContainerDied","Data":"a054eac97e88cce1aa9882153236f28817412a343a3608ef402b154e16e8b759"} Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.134758 4948 generic.go:334] "Generic (PLEG): container finished" podID="89b49c50-bae7-47fe-b85f-7a718953933d" containerID="27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9" exitCode=0 Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.134789 4948 generic.go:334] "Generic (PLEG): container finished" podID="89b49c50-bae7-47fe-b85f-7a718953933d" containerID="61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07" exitCode=2 Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.134797 4948 generic.go:334] "Generic (PLEG): container finished" podID="89b49c50-bae7-47fe-b85f-7a718953933d" containerID="f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56" exitCode=0 Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.134815 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerDied","Data":"27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9"} Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.134841 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerDied","Data":"61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07"} Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.134852 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerDied","Data":"f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56"} Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.167180 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6ffd468-xgk27"] Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.362083 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5dgw7"] Feb 20 08:23:07 crc kubenswrapper[4948]: W0220 08:23:07.369658 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63124eef_9456_44d1_a8ba_cc8297af3e9b.slice/crio-b02b398d101ad6f54b2deb0825184d401d1e5b0ad4e38257f4120a442c5a5509 WatchSource:0}: Error finding container b02b398d101ad6f54b2deb0825184d401d1e5b0ad4e38257f4120a442c5a5509: Status 404 returned error can't find the container with id b02b398d101ad6f54b2deb0825184d401d1e5b0ad4e38257f4120a442c5a5509 Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.412307 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65487cc4d6-7gdvr"] Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.433301 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.584173 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-swift-storage-0\") pod \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.584272 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6wdv\" (UniqueName: \"kubernetes.io/projected/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-kube-api-access-c6wdv\") pod \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.584317 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-svc\") pod \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.584348 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-config\") pod \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.584389 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-sb\") pod \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.584416 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-nb\") pod \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\" (UID: \"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e\") " Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.596175 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-kube-api-access-c6wdv" (OuterVolumeSpecName: "kube-api-access-c6wdv") pod "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" (UID: "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e"). InnerVolumeSpecName "kube-api-access-c6wdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.625576 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" (UID: "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.635254 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" (UID: "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.637472 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-config" (OuterVolumeSpecName: "config") pod "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" (UID: "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.638838 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" (UID: "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.639942 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" (UID: "4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.688433 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.688513 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6wdv\" (UniqueName: \"kubernetes.io/projected/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-kube-api-access-c6wdv\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.688536 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.688550 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.688600 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.688613 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:07 crc kubenswrapper[4948]: I0220 08:23:07.888225 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.049995 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.144909 4948 generic.go:334] "Generic (PLEG): container finished" podID="63124eef-9456-44d1-a8ba-cc8297af3e9b" containerID="56e628ea61223b6bdd0b21738d110aaaa403a55ddd583614f61bc57a6cdb16fd" exitCode=0 Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.145023 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" event={"ID":"63124eef-9456-44d1-a8ba-cc8297af3e9b","Type":"ContainerDied","Data":"56e628ea61223b6bdd0b21738d110aaaa403a55ddd583614f61bc57a6cdb16fd"} Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.145087 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" event={"ID":"63124eef-9456-44d1-a8ba-cc8297af3e9b","Type":"ContainerStarted","Data":"b02b398d101ad6f54b2deb0825184d401d1e5b0ad4e38257f4120a442c5a5509"} Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.149984 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1ff6993b-2bde-429d-b0c7-cdea51f733a2","Type":"ContainerStarted","Data":"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d"} Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.156378 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6ffd468-xgk27" event={"ID":"292d8e1e-de3b-4259-afef-c19e2c21187c","Type":"ContainerStarted","Data":"f6496764c718a0f648853d16537d4d7261c22048971ad4db61bf7b138d835187"} Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.157769 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"75294aee-c7f3-4b0b-82fc-df058490b90f","Type":"ContainerStarted","Data":"0c8a24ef5e8c2ed45ddfc6908331dbda28dd167039cb8bc7f442c18afef4d964"} Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.161089 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65487cc4d6-7gdvr" event={"ID":"6a669dd0-d287-4d07-9eec-aea5ed80c791","Type":"ContainerStarted","Data":"86e858c136de80a807f997b8991e918a47db690a2c7b9fcc3401caa707f989dd"} Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.161150 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65487cc4d6-7gdvr" event={"ID":"6a669dd0-d287-4d07-9eec-aea5ed80c791","Type":"ContainerStarted","Data":"b3b840325250a810c158ca1f61d60db618feec352dbf68e16aad14e15139327d"} Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.168939 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-97gsc" event={"ID":"4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e","Type":"ContainerDied","Data":"a73968f29763e85cb2562473f92370ee9d88e9ecbb9c1dd9bc348a65b2728c01"} Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.168996 4948 scope.go:117] "RemoveContainer" containerID="a054eac97e88cce1aa9882153236f28817412a343a3608ef402b154e16e8b759" Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.169146 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-97gsc" Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.388034 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-97gsc"] Feb 20 08:23:08 crc kubenswrapper[4948]: I0220 08:23:08.401235 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-97gsc"] Feb 20 08:23:09 crc kubenswrapper[4948]: I0220 08:23:09.429695 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:23:09 crc kubenswrapper[4948]: I0220 08:23:09.526541 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:23:09 crc kubenswrapper[4948]: I0220 08:23:09.666147 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hh5s2"] Feb 20 08:23:09 crc kubenswrapper[4948]: I0220 08:23:09.738320 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" path="/var/lib/kubelet/pods/4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e/volumes" Feb 20 08:23:09 crc kubenswrapper[4948]: I0220 08:23:09.965399 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-685c4db87b-rw57h" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.059477 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66b4d4cb8-hgmkd"] Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.059724 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66b4d4cb8-hgmkd" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon-log" containerID="cri-o://577865ca69f0d8457d617948e5cac2be0a124e4620edc60df1395130de8a2334" gracePeriod=30 Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.059783 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66b4d4cb8-hgmkd" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon" containerID="cri-o://a1b60a312d88b84db5225c1aff5290457499b2775b6fffec7ca3246896b28523" gracePeriod=30 Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.190084 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1ff6993b-2bde-429d-b0c7-cdea51f733a2","Type":"ContainerStarted","Data":"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3"} Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.190170 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.190159 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerName="cinder-api-log" containerID="cri-o://2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d" gracePeriod=30 Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.190218 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerName="cinder-api" containerID="cri-o://0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3" gracePeriod=30 Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.194571 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"75294aee-c7f3-4b0b-82fc-df058490b90f","Type":"ContainerStarted","Data":"c1fbafc4d60f4668a3ce40442dbec31bba82bcb8754ebb88a8aed726953af8e7"} Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.200753 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65487cc4d6-7gdvr" event={"ID":"6a669dd0-d287-4d07-9eec-aea5ed80c791","Type":"ContainerStarted","Data":"ab05c86cbd431ad2cc4a3cef37a6fc481837d1b6073c9a80ed0ee92d983240c8"} Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.200986 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.202801 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" event={"ID":"63124eef-9456-44d1-a8ba-cc8297af3e9b","Type":"ContainerStarted","Data":"724a55a11670e45fe3d2ce7a82832ba87e341cd787eee57c186dd30931fb8402"} Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.212951 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.212936127 podStartE2EDuration="5.212936127s" podCreationTimestamp="2026-02-20 08:23:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:10.2065656 +0000 UTC m=+1039.181060430" watchObservedRunningTime="2026-02-20 08:23:10.212936127 +0000 UTC m=+1039.187430947" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.234557 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.472942855 podStartE2EDuration="5.234537932s" podCreationTimestamp="2026-02-20 08:23:05 +0000 UTC" firstStartedPulling="2026-02-20 08:23:06.127056574 +0000 UTC m=+1035.101551394" lastFinishedPulling="2026-02-20 08:23:06.888651651 +0000 UTC m=+1035.863146471" observedRunningTime="2026-02-20 08:23:10.231069406 +0000 UTC m=+1039.205564226" watchObservedRunningTime="2026-02-20 08:23:10.234537932 +0000 UTC m=+1039.209032752" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.257947 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-65487cc4d6-7gdvr" podStartSLOduration=4.25792403 podStartE2EDuration="4.25792403s" podCreationTimestamp="2026-02-20 08:23:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:10.247431931 +0000 UTC m=+1039.221926761" watchObservedRunningTime="2026-02-20 08:23:10.25792403 +0000 UTC m=+1039.232418850" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.271674 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" podStartSLOduration=4.27165913 podStartE2EDuration="4.27165913s" podCreationTimestamp="2026-02-20 08:23:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:10.268725188 +0000 UTC m=+1039.243220008" watchObservedRunningTime="2026-02-20 08:23:10.27165913 +0000 UTC m=+1039.246153950" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.562881 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.804269 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.870126 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.958949 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-scripts\") pod \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959364 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xq6j4\" (UniqueName: \"kubernetes.io/projected/1ff6993b-2bde-429d-b0c7-cdea51f733a2-kube-api-access-xq6j4\") pod \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959385 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ff6993b-2bde-429d-b0c7-cdea51f733a2-etc-machine-id\") pod \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959404 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-combined-ca-bundle\") pod \"89b49c50-bae7-47fe-b85f-7a718953933d\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959475 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77njn\" (UniqueName: \"kubernetes.io/projected/89b49c50-bae7-47fe-b85f-7a718953933d-kube-api-access-77njn\") pod \"89b49c50-bae7-47fe-b85f-7a718953933d\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959500 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-log-httpd\") pod \"89b49c50-bae7-47fe-b85f-7a718953933d\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959521 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-scripts\") pod \"89b49c50-bae7-47fe-b85f-7a718953933d\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959552 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ff6993b-2bde-429d-b0c7-cdea51f733a2-logs\") pod \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959573 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-config-data\") pod \"89b49c50-bae7-47fe-b85f-7a718953933d\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959582 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ff6993b-2bde-429d-b0c7-cdea51f733a2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1ff6993b-2bde-429d-b0c7-cdea51f733a2" (UID: "1ff6993b-2bde-429d-b0c7-cdea51f733a2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959607 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-combined-ca-bundle\") pod \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959734 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-sg-core-conf-yaml\") pod \"89b49c50-bae7-47fe-b85f-7a718953933d\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959790 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-run-httpd\") pod \"89b49c50-bae7-47fe-b85f-7a718953933d\" (UID: \"89b49c50-bae7-47fe-b85f-7a718953933d\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959815 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data\") pod \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.959843 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data-custom\") pod \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\" (UID: \"1ff6993b-2bde-429d-b0c7-cdea51f733a2\") " Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.960274 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ff6993b-2bde-429d-b0c7-cdea51f733a2-logs" (OuterVolumeSpecName: "logs") pod "1ff6993b-2bde-429d-b0c7-cdea51f733a2" (UID: "1ff6993b-2bde-429d-b0c7-cdea51f733a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.960299 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "89b49c50-bae7-47fe-b85f-7a718953933d" (UID: "89b49c50-bae7-47fe-b85f-7a718953933d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.960735 4948 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.960757 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ff6993b-2bde-429d-b0c7-cdea51f733a2-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.960901 4948 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ff6993b-2bde-429d-b0c7-cdea51f733a2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.960934 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "89b49c50-bae7-47fe-b85f-7a718953933d" (UID: "89b49c50-bae7-47fe-b85f-7a718953933d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.963770 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89b49c50-bae7-47fe-b85f-7a718953933d-kube-api-access-77njn" (OuterVolumeSpecName: "kube-api-access-77njn") pod "89b49c50-bae7-47fe-b85f-7a718953933d" (UID: "89b49c50-bae7-47fe-b85f-7a718953933d"). InnerVolumeSpecName "kube-api-access-77njn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.966717 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-scripts" (OuterVolumeSpecName: "scripts") pod "89b49c50-bae7-47fe-b85f-7a718953933d" (UID: "89b49c50-bae7-47fe-b85f-7a718953933d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.969392 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-scripts" (OuterVolumeSpecName: "scripts") pod "1ff6993b-2bde-429d-b0c7-cdea51f733a2" (UID: "1ff6993b-2bde-429d-b0c7-cdea51f733a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.969417 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ff6993b-2bde-429d-b0c7-cdea51f733a2-kube-api-access-xq6j4" (OuterVolumeSpecName: "kube-api-access-xq6j4") pod "1ff6993b-2bde-429d-b0c7-cdea51f733a2" (UID: "1ff6993b-2bde-429d-b0c7-cdea51f733a2"). InnerVolumeSpecName "kube-api-access-xq6j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.970223 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1ff6993b-2bde-429d-b0c7-cdea51f733a2" (UID: "1ff6993b-2bde-429d-b0c7-cdea51f733a2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.992370 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ff6993b-2bde-429d-b0c7-cdea51f733a2" (UID: "1ff6993b-2bde-429d-b0c7-cdea51f733a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:10 crc kubenswrapper[4948]: I0220 08:23:10.992741 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "89b49c50-bae7-47fe-b85f-7a718953933d" (UID: "89b49c50-bae7-47fe-b85f-7a718953933d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.016496 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data" (OuterVolumeSpecName: "config-data") pod "1ff6993b-2bde-429d-b0c7-cdea51f733a2" (UID: "1ff6993b-2bde-429d-b0c7-cdea51f733a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.038190 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89b49c50-bae7-47fe-b85f-7a718953933d" (UID: "89b49c50-bae7-47fe-b85f-7a718953933d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.061765 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-config-data" (OuterVolumeSpecName: "config-data") pod "89b49c50-bae7-47fe-b85f-7a718953933d" (UID: "89b49c50-bae7-47fe-b85f-7a718953933d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062135 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062160 4948 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062171 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062179 4948 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89b49c50-bae7-47fe-b85f-7a718953933d-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062188 4948 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062195 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ff6993b-2bde-429d-b0c7-cdea51f733a2-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062205 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xq6j4\" (UniqueName: \"kubernetes.io/projected/1ff6993b-2bde-429d-b0c7-cdea51f733a2-kube-api-access-xq6j4\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062215 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062223 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77njn\" (UniqueName: \"kubernetes.io/projected/89b49c50-bae7-47fe-b85f-7a718953933d-kube-api-access-77njn\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062231 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.062238 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b49c50-bae7-47fe-b85f-7a718953933d-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.216274 4948 generic.go:334] "Generic (PLEG): container finished" podID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerID="a1b60a312d88b84db5225c1aff5290457499b2775b6fffec7ca3246896b28523" exitCode=0 Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.216362 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66b4d4cb8-hgmkd" event={"ID":"d1046503-c6d3-49bf-9523-2f897fcb82ba","Type":"ContainerDied","Data":"a1b60a312d88b84db5225c1aff5290457499b2775b6fffec7ca3246896b28523"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.224249 4948 generic.go:334] "Generic (PLEG): container finished" podID="89b49c50-bae7-47fe-b85f-7a718953933d" containerID="68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493" exitCode=0 Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.224283 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerDied","Data":"68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.224336 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89b49c50-bae7-47fe-b85f-7a718953933d","Type":"ContainerDied","Data":"2c29cb12a24e0769e2fc11186d9dff0648d1bcb261fcbab7060faff8fb76085a"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.224347 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.224358 4948 scope.go:117] "RemoveContainer" containerID="27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.227345 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76d486c65f-w9lc6" event={"ID":"4b356058-d957-4fa5-a106-3fe5052c8539","Type":"ContainerStarted","Data":"429d83e2cd3914efc5db9d90e28d0b14b09e14a603ac3bc8a797688781218235"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.227385 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76d486c65f-w9lc6" event={"ID":"4b356058-d957-4fa5-a106-3fe5052c8539","Type":"ContainerStarted","Data":"97fe5c50373074a6e2530e57e930ec6666df51919433a7f7204067aad9e98819"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.231553 4948 generic.go:334] "Generic (PLEG): container finished" podID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerID="0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3" exitCode=0 Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.231593 4948 generic.go:334] "Generic (PLEG): container finished" podID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerID="2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d" exitCode=143 Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.231639 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1ff6993b-2bde-429d-b0c7-cdea51f733a2","Type":"ContainerDied","Data":"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.231666 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1ff6993b-2bde-429d-b0c7-cdea51f733a2","Type":"ContainerDied","Data":"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.231684 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1ff6993b-2bde-429d-b0c7-cdea51f733a2","Type":"ContainerDied","Data":"b0f8def01191a21afa9e066696a67301fbaaf496e33dd51ae806612c580bc882"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.231754 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.251411 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6ffd468-xgk27" event={"ID":"292d8e1e-de3b-4259-afef-c19e2c21187c","Type":"ContainerStarted","Data":"06bcf0b4075319148a17c9d445b9a694bf41ad80dfe19d1b6d5b8f09c4a2b65f"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.251623 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6ffd468-xgk27" event={"ID":"292d8e1e-de3b-4259-afef-c19e2c21187c","Type":"ContainerStarted","Data":"33432f8fb31556de3b070d1c39482bcc6d6666ff0bb7281f5a085f52c18b5f5d"} Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.252457 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.252508 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.253532 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hh5s2" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="registry-server" containerID="cri-o://3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6" gracePeriod=2 Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.272744 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-76d486c65f-w9lc6" podStartSLOduration=1.9453027710000002 podStartE2EDuration="5.272717884s" podCreationTimestamp="2026-02-20 08:23:06 +0000 UTC" firstStartedPulling="2026-02-20 08:23:07.095361807 +0000 UTC m=+1036.069856627" lastFinishedPulling="2026-02-20 08:23:10.42277693 +0000 UTC m=+1039.397271740" observedRunningTime="2026-02-20 08:23:11.247873779 +0000 UTC m=+1040.222368629" watchObservedRunningTime="2026-02-20 08:23:11.272717884 +0000 UTC m=+1040.247212714" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.309301 4948 scope.go:117] "RemoveContainer" containerID="61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.325924 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6ffd468-xgk27" podStartSLOduration=2.061738272 podStartE2EDuration="5.32589663s" podCreationTimestamp="2026-02-20 08:23:06 +0000 UTC" firstStartedPulling="2026-02-20 08:23:07.169285536 +0000 UTC m=+1036.143780356" lastFinishedPulling="2026-02-20 08:23:10.433443894 +0000 UTC m=+1039.407938714" observedRunningTime="2026-02-20 08:23:11.279224405 +0000 UTC m=+1040.253719265" watchObservedRunningTime="2026-02-20 08:23:11.32589663 +0000 UTC m=+1040.300391450" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.352575 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.358845 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.412135 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.433454 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.433938 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="sg-core" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.433960 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="sg-core" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.433992 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="ceilometer-notification-agent" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434001 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="ceilometer-notification-agent" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.434020 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" containerName="init" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434028 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" containerName="init" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.434043 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="ceilometer-central-agent" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434050 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="ceilometer-central-agent" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.434069 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="proxy-httpd" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434075 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="proxy-httpd" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.434099 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerName="cinder-api" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434107 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerName="cinder-api" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.434124 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerName="cinder-api-log" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434131 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerName="cinder-api-log" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434329 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="ceilometer-central-agent" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434343 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="sg-core" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434358 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e601cb0-67e4-4cc4-b7e8-41f1dfa9966e" containerName="init" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434375 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="proxy-httpd" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434388 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerName="cinder-api-log" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434402 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" containerName="cinder-api" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.434417 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" containerName="ceilometer-notification-agent" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.436368 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.440350 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.443125 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.477142 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.505455 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.546057 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.547983 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.551823 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.552082 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.552301 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.557029 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.567242 4948 scope.go:117] "RemoveContainer" containerID="68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.580538 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-scripts\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.580597 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-config-data\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.580739 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.580954 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rkd9\" (UniqueName: \"kubernetes.io/projected/44092649-be82-4225-8cfc-d37dbe840877-kube-api-access-4rkd9\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.581203 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.581353 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-run-httpd\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.581524 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-log-httpd\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.622137 4948 scope.go:117] "RemoveContainer" containerID="f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.659076 4948 scope.go:117] "RemoveContainer" containerID="27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.659577 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9\": container with ID starting with 27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9 not found: ID does not exist" containerID="27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.659623 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9"} err="failed to get container status \"27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9\": rpc error: code = NotFound desc = could not find container \"27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9\": container with ID starting with 27fa5fd7b38b4d8b359bee17f753923040dddd64a704b2cd2a15217dcb4497e9 not found: ID does not exist" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.659651 4948 scope.go:117] "RemoveContainer" containerID="61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.663061 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07\": container with ID starting with 61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07 not found: ID does not exist" containerID="61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.663101 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07"} err="failed to get container status \"61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07\": rpc error: code = NotFound desc = could not find container \"61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07\": container with ID starting with 61ecb15080679db6c7523b2b45faf0110c6e8bc826d5ac4d37af1077d531ed07 not found: ID does not exist" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.663127 4948 scope.go:117] "RemoveContainer" containerID="68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.663496 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493\": container with ID starting with 68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493 not found: ID does not exist" containerID="68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.663522 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493"} err="failed to get container status \"68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493\": rpc error: code = NotFound desc = could not find container \"68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493\": container with ID starting with 68df1bb1b13a5e62c745d68de5298b184c0cdcaa037a022eedeb800b70856493 not found: ID does not exist" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.663538 4948 scope.go:117] "RemoveContainer" containerID="f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.668160 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56\": container with ID starting with f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56 not found: ID does not exist" containerID="f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.668200 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56"} err="failed to get container status \"f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56\": rpc error: code = NotFound desc = could not find container \"f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56\": container with ID starting with f21ea8eddb7fd63986bb5357a809bc7c94ba2e30ea5976b9795bdbd7d103ba56 not found: ID does not exist" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.668225 4948 scope.go:117] "RemoveContainer" containerID="0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.683841 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-run-httpd\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.683887 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-scripts\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.683928 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-log-httpd\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.683956 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-config-data-custom\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.683988 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcn5v\" (UniqueName: \"kubernetes.io/projected/e5967cab-cc29-4c87-b873-0f5d20203d4f-kube-api-access-mcn5v\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684005 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-scripts\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684027 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-config-data\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684046 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-config-data\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684061 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684091 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e5967cab-cc29-4c87-b873-0f5d20203d4f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684118 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rkd9\" (UniqueName: \"kubernetes.io/projected/44092649-be82-4225-8cfc-d37dbe840877-kube-api-access-4rkd9\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684145 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684173 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5967cab-cc29-4c87-b873-0f5d20203d4f-logs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684192 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684215 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684235 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684659 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-run-httpd\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.684881 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-log-httpd\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.685999 4948 scope.go:117] "RemoveContainer" containerID="2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.691602 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.691941 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-scripts\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.693492 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-config-data\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.708480 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.716770 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rkd9\" (UniqueName: \"kubernetes.io/projected/44092649-be82-4225-8cfc-d37dbe840877-kube-api-access-4rkd9\") pod \"ceilometer-0\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.741322 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ff6993b-2bde-429d-b0c7-cdea51f733a2" path="/var/lib/kubelet/pods/1ff6993b-2bde-429d-b0c7-cdea51f733a2/volumes" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.742076 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89b49c50-bae7-47fe-b85f-7a718953933d" path="/var/lib/kubelet/pods/89b49c50-bae7-47fe-b85f-7a718953933d/volumes" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.781485 4948 scope.go:117] "RemoveContainer" containerID="0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.781967 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3\": container with ID starting with 0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3 not found: ID does not exist" containerID="0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.782030 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3"} err="failed to get container status \"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3\": rpc error: code = NotFound desc = could not find container \"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3\": container with ID starting with 0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3 not found: ID does not exist" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.782072 4948 scope.go:117] "RemoveContainer" containerID="2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d" Feb 20 08:23:11 crc kubenswrapper[4948]: E0220 08:23:11.782556 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d\": container with ID starting with 2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d not found: ID does not exist" containerID="2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.782596 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d"} err="failed to get container status \"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d\": rpc error: code = NotFound desc = could not find container \"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d\": container with ID starting with 2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d not found: ID does not exist" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.782628 4948 scope.go:117] "RemoveContainer" containerID="0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.783153 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3"} err="failed to get container status \"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3\": rpc error: code = NotFound desc = could not find container \"0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3\": container with ID starting with 0f2c76fb7b2272891c92989048ced16e6a9de109d58761f70e320c1c141107d3 not found: ID does not exist" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.783187 4948 scope.go:117] "RemoveContainer" containerID="2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.783485 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d"} err="failed to get container status \"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d\": rpc error: code = NotFound desc = could not find container \"2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d\": container with ID starting with 2cf89059747c86a7611df4842eeab89fcfa4211ffed1cd47ac35057761460d3d not found: ID does not exist" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.785682 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-config-data-custom\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.785752 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcn5v\" (UniqueName: \"kubernetes.io/projected/e5967cab-cc29-4c87-b873-0f5d20203d4f-kube-api-access-mcn5v\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.786320 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-config-data\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.787161 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e5967cab-cc29-4c87-b873-0f5d20203d4f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.787238 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e5967cab-cc29-4c87-b873-0f5d20203d4f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.787317 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.787850 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5967cab-cc29-4c87-b873-0f5d20203d4f-logs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.787889 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.787922 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.788025 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-scripts\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.788325 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5967cab-cc29-4c87-b873-0f5d20203d4f-logs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.789197 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-config-data-custom\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.790853 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-config-data\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.791870 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.792592 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.793187 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.794289 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5967cab-cc29-4c87-b873-0f5d20203d4f-scripts\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.802225 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcn5v\" (UniqueName: \"kubernetes.io/projected/e5967cab-cc29-4c87-b873-0f5d20203d4f-kube-api-access-mcn5v\") pod \"cinder-api-0\" (UID: \"e5967cab-cc29-4c87-b873-0f5d20203d4f\") " pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.846563 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.865965 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.889493 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.936224 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.992048 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-utilities\") pod \"426b37f9-4acb-4d28-a55c-3fcde2600002\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.992323 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-catalog-content\") pod \"426b37f9-4acb-4d28-a55c-3fcde2600002\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.992428 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw6bb\" (UniqueName: \"kubernetes.io/projected/426b37f9-4acb-4d28-a55c-3fcde2600002-kube-api-access-tw6bb\") pod \"426b37f9-4acb-4d28-a55c-3fcde2600002\" (UID: \"426b37f9-4acb-4d28-a55c-3fcde2600002\") " Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.992793 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-utilities" (OuterVolumeSpecName: "utilities") pod "426b37f9-4acb-4d28-a55c-3fcde2600002" (UID: "426b37f9-4acb-4d28-a55c-3fcde2600002"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:11 crc kubenswrapper[4948]: I0220 08:23:11.997281 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/426b37f9-4acb-4d28-a55c-3fcde2600002-kube-api-access-tw6bb" (OuterVolumeSpecName: "kube-api-access-tw6bb") pod "426b37f9-4acb-4d28-a55c-3fcde2600002" (UID: "426b37f9-4acb-4d28-a55c-3fcde2600002"). InnerVolumeSpecName "kube-api-access-tw6bb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.079572 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "426b37f9-4acb-4d28-a55c-3fcde2600002" (UID: "426b37f9-4acb-4d28-a55c-3fcde2600002"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.096477 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.096538 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw6bb\" (UniqueName: \"kubernetes.io/projected/426b37f9-4acb-4d28-a55c-3fcde2600002-kube-api-access-tw6bb\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.096550 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/426b37f9-4acb-4d28-a55c-3fcde2600002-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.136798 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-686d666564-8tnjb"] Feb 20 08:23:12 crc kubenswrapper[4948]: E0220 08:23:12.137395 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="registry-server" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.137410 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="registry-server" Feb 20 08:23:12 crc kubenswrapper[4948]: E0220 08:23:12.137421 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="extract-utilities" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.137427 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="extract-utilities" Feb 20 08:23:12 crc kubenswrapper[4948]: E0220 08:23:12.137447 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="extract-content" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.137453 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="extract-content" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.137607 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerName="registry-server" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.138479 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.141797 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.143101 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.170835 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-686d666564-8tnjb"] Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.199026 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-internal-tls-certs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.199103 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-config-data-custom\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.199156 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-public-tls-certs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.199205 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-config-data\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.199227 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51f6301c-57be-4a07-91df-b1a9f90e3f40-logs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.199255 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-combined-ca-bundle\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.199274 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fdqm\" (UniqueName: \"kubernetes.io/projected/51f6301c-57be-4a07-91df-b1a9f90e3f40-kube-api-access-5fdqm\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.292240 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-685bb7f5c-jd9nm"] Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.292465 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-685bb7f5c-jd9nm" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-api" containerID="cri-o://1ad25f5bfef4b0fa06a939cc9fb8e5bfa3dad488d3265a3cd5fa655c097dda5b" gracePeriod=30 Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.292819 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-685bb7f5c-jd9nm" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-httpd" containerID="cri-o://d2ca1451e75be7daafc6907a8305baa229ed6441f1f79895bd2df00e5c17bfa7" gracePeriod=30 Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.300838 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-public-tls-certs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.300937 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-config-data\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.300982 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51f6301c-57be-4a07-91df-b1a9f90e3f40-logs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.301030 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-combined-ca-bundle\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.301056 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fdqm\" (UniqueName: \"kubernetes.io/projected/51f6301c-57be-4a07-91df-b1a9f90e3f40-kube-api-access-5fdqm\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.301110 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-internal-tls-certs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.301167 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-config-data-custom\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.301574 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51f6301c-57be-4a07-91df-b1a9f90e3f40-logs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.305303 4948 generic.go:334] "Generic (PLEG): container finished" podID="426b37f9-4acb-4d28-a55c-3fcde2600002" containerID="3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6" exitCode=0 Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.308367 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hh5s2" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.309230 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh5s2" event={"ID":"426b37f9-4acb-4d28-a55c-3fcde2600002","Type":"ContainerDied","Data":"3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6"} Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.309268 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hh5s2" event={"ID":"426b37f9-4acb-4d28-a55c-3fcde2600002","Type":"ContainerDied","Data":"6c8a403f4e4cb34b74e5da72f8bdf18b9a2a922052dce8500240a25dba684641"} Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.309286 4948 scope.go:117] "RemoveContainer" containerID="3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.317187 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-config-data\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.319653 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-public-tls-certs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.323265 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-685bb7f5c-jd9nm" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.156:9696/\": EOF" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.326784 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-config-data-custom\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.336929 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fdqm\" (UniqueName: \"kubernetes.io/projected/51f6301c-57be-4a07-91df-b1a9f90e3f40-kube-api-access-5fdqm\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.342495 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-combined-ca-bundle\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.358918 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-76b5684765-h5624"] Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.360330 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.364297 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51f6301c-57be-4a07-91df-b1a9f90e3f40-internal-tls-certs\") pod \"barbican-api-686d666564-8tnjb\" (UID: \"51f6301c-57be-4a07-91df-b1a9f90e3f40\") " pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.379083 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-76b5684765-h5624"] Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.443895 4948 scope.go:117] "RemoveContainer" containerID="2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.444021 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.454767 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hh5s2"] Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.462664 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hh5s2"] Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.465925 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:12 crc kubenswrapper[4948]: W0220 08:23:12.504113 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5967cab_cc29_4c87_b873_0f5d20203d4f.slice/crio-29185c36a66efe969c66c1ca5274cb67c9aad81d503060845b8062dd34d95951 WatchSource:0}: Error finding container 29185c36a66efe969c66c1ca5274cb67c9aad81d503060845b8062dd34d95951: Status 404 returned error can't find the container with id 29185c36a66efe969c66c1ca5274cb67c9aad81d503060845b8062dd34d95951 Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.510171 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-internal-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.510359 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-ovndb-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.510575 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-httpd-config\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.510698 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-combined-ca-bundle\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.510933 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-public-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.511085 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f67zh\" (UniqueName: \"kubernetes.io/projected/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-kube-api-access-f67zh\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.511265 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-config\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.512361 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.582854 4948 scope.go:117] "RemoveContainer" containerID="62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.613017 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-config\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.613099 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-internal-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.613119 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-ovndb-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.613215 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-httpd-config\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.613258 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-combined-ca-bundle\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.613275 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-public-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.613328 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f67zh\" (UniqueName: \"kubernetes.io/projected/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-kube-api-access-f67zh\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.618301 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-ovndb-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.618390 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-config\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.622753 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-combined-ca-bundle\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.627635 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-public-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.629818 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-httpd-config\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.630508 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-internal-tls-certs\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.638591 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f67zh\" (UniqueName: \"kubernetes.io/projected/fb9d5a3e-b07f-4229-9b5b-d0db6750cf59-kube-api-access-f67zh\") pod \"neutron-76b5684765-h5624\" (UID: \"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59\") " pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.639104 4948 scope.go:117] "RemoveContainer" containerID="3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6" Feb 20 08:23:12 crc kubenswrapper[4948]: E0220 08:23:12.639493 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6\": container with ID starting with 3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6 not found: ID does not exist" containerID="3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.639575 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6"} err="failed to get container status \"3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6\": rpc error: code = NotFound desc = could not find container \"3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6\": container with ID starting with 3d2681901d856e71e99e2f5dd2205106a0fb9176c61fcdad6f85c78cea4118e6 not found: ID does not exist" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.639644 4948 scope.go:117] "RemoveContainer" containerID="2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d" Feb 20 08:23:12 crc kubenswrapper[4948]: E0220 08:23:12.639898 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d\": container with ID starting with 2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d not found: ID does not exist" containerID="2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.639982 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d"} err="failed to get container status \"2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d\": rpc error: code = NotFound desc = could not find container \"2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d\": container with ID starting with 2d58a1beb2801a36e08660bfb99639b7abb52f4d18887bdac3f68efb75c8316d not found: ID does not exist" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.640053 4948 scope.go:117] "RemoveContainer" containerID="62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863" Feb 20 08:23:12 crc kubenswrapper[4948]: E0220 08:23:12.644587 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863\": container with ID starting with 62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863 not found: ID does not exist" containerID="62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.644635 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863"} err="failed to get container status \"62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863\": rpc error: code = NotFound desc = could not find container \"62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863\": container with ID starting with 62e9964b1041b8a3ee083b7749dd1ff49cbef0514cc265f8b36923ced11f9863 not found: ID does not exist" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.740303 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:12 crc kubenswrapper[4948]: I0220 08:23:12.940486 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-686d666564-8tnjb"] Feb 20 08:23:12 crc kubenswrapper[4948]: W0220 08:23:12.966317 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51f6301c_57be_4a07_91df_b1a9f90e3f40.slice/crio-af0403aeb9d6e04d55a315e4c351060deae309bea1ede2db1e668df78c7110a7 WatchSource:0}: Error finding container af0403aeb9d6e04d55a315e4c351060deae309bea1ede2db1e668df78c7110a7: Status 404 returned error can't find the container with id af0403aeb9d6e04d55a315e4c351060deae309bea1ede2db1e668df78c7110a7 Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.282799 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-76b5684765-h5624"] Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.321411 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-686d666564-8tnjb" event={"ID":"51f6301c-57be-4a07-91df-b1a9f90e3f40","Type":"ContainerStarted","Data":"8d925a9cf348b2c3c2fa4ee34cb7d0ae0a9d07d119fd8e257a3a9e7f5a687bbd"} Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.321453 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-686d666564-8tnjb" event={"ID":"51f6301c-57be-4a07-91df-b1a9f90e3f40","Type":"ContainerStarted","Data":"af0403aeb9d6e04d55a315e4c351060deae309bea1ede2db1e668df78c7110a7"} Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.325234 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76b5684765-h5624" event={"ID":"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59","Type":"ContainerStarted","Data":"38aa7807e10bebc2e535552cb2dda45a3e1972a20649904b4402700bc8d8dd22"} Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.328032 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerStarted","Data":"eab93963947d8aecd2b45387be911464a6576e8236c9ccd3d0a009ad65054a74"} Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.328150 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerStarted","Data":"8266937b599da958a185777fc1969453e86491578f18e8d8f35aa2e0681bba26"} Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.329317 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e5967cab-cc29-4c87-b873-0f5d20203d4f","Type":"ContainerStarted","Data":"af195efb8dccb5b78426393f4467711f39f97b1e3850500984806b51282213f6"} Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.329451 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e5967cab-cc29-4c87-b873-0f5d20203d4f","Type":"ContainerStarted","Data":"29185c36a66efe969c66c1ca5274cb67c9aad81d503060845b8062dd34d95951"} Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.370949 4948 generic.go:334] "Generic (PLEG): container finished" podID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerID="d2ca1451e75be7daafc6907a8305baa229ed6441f1f79895bd2df00e5c17bfa7" exitCode=0 Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.371178 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685bb7f5c-jd9nm" event={"ID":"e0661780-ca2b-42ae-be25-5d0e914fbb49","Type":"ContainerDied","Data":"d2ca1451e75be7daafc6907a8305baa229ed6441f1f79895bd2df00e5c17bfa7"} Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.737440 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="426b37f9-4acb-4d28-a55c-3fcde2600002" path="/var/lib/kubelet/pods/426b37f9-4acb-4d28-a55c-3fcde2600002/volumes" Feb 20 08:23:13 crc kubenswrapper[4948]: I0220 08:23:13.927583 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-685bb7f5c-jd9nm" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.156:9696/\": dial tcp 10.217.0.156:9696: connect: connection refused" Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.067568 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.379005 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e5967cab-cc29-4c87-b873-0f5d20203d4f","Type":"ContainerStarted","Data":"e49db6d27fc81d3422ce0d425640ec2f5701bf916f725129edc6e1101abe2979"} Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.380015 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.382237 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-686d666564-8tnjb" event={"ID":"51f6301c-57be-4a07-91df-b1a9f90e3f40","Type":"ContainerStarted","Data":"9d42d5115f0ed851367331fc56d5ac48b9ae3f6d9455032d65b1f2b001c75c79"} Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.382600 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.384265 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76b5684765-h5624" event={"ID":"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59","Type":"ContainerStarted","Data":"09aa54f86c858f9fc51ae59c1879f3a7cbc7a7847eacf1ae0b7ae4f33b0273fc"} Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.384311 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76b5684765-h5624" event={"ID":"fb9d5a3e-b07f-4229-9b5b-d0db6750cf59","Type":"ContainerStarted","Data":"2ae99d6a857d168ec7f201c1bbf30b4bd88e112bcf930306d6a6abfbab050ad0"} Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.384466 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.386464 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerStarted","Data":"7e6d02d6234c2e72678f463fac466a22ef0548ef1c161d0d356079f01463ff09"} Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.413582 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.41355432 podStartE2EDuration="3.41355432s" podCreationTimestamp="2026-02-20 08:23:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:14.406642149 +0000 UTC m=+1043.381136969" watchObservedRunningTime="2026-02-20 08:23:14.41355432 +0000 UTC m=+1043.388049140" Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.437961 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-686d666564-8tnjb" podStartSLOduration=2.437940353 podStartE2EDuration="2.437940353s" podCreationTimestamp="2026-02-20 08:23:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:14.430391997 +0000 UTC m=+1043.404886817" watchObservedRunningTime="2026-02-20 08:23:14.437940353 +0000 UTC m=+1043.412435173" Feb 20 08:23:14 crc kubenswrapper[4948]: I0220 08:23:14.461361 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-76b5684765-h5624" podStartSLOduration=2.461341902 podStartE2EDuration="2.461341902s" podCreationTimestamp="2026-02-20 08:23:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:14.45639865 +0000 UTC m=+1043.430893470" watchObservedRunningTime="2026-02-20 08:23:14.461341902 +0000 UTC m=+1043.435836722" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.396959 4948 generic.go:334] "Generic (PLEG): container finished" podID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerID="1ad25f5bfef4b0fa06a939cc9fb8e5bfa3dad488d3265a3cd5fa655c097dda5b" exitCode=0 Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.397050 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685bb7f5c-jd9nm" event={"ID":"e0661780-ca2b-42ae-be25-5d0e914fbb49","Type":"ContainerDied","Data":"1ad25f5bfef4b0fa06a939cc9fb8e5bfa3dad488d3265a3cd5fa655c097dda5b"} Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.401482 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerStarted","Data":"d829cc037c7ba0bd0abfe863863495a3388592344985fae656879df8c40ec990"} Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.401912 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.555428 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.766702 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.811740 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.846679 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.876568 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-config\") pod \"e0661780-ca2b-42ae-be25-5d0e914fbb49\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.876629 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-internal-tls-certs\") pod \"e0661780-ca2b-42ae-be25-5d0e914fbb49\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.876677 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-combined-ca-bundle\") pod \"e0661780-ca2b-42ae-be25-5d0e914fbb49\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.876732 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-public-tls-certs\") pod \"e0661780-ca2b-42ae-be25-5d0e914fbb49\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.876779 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-ovndb-tls-certs\") pod \"e0661780-ca2b-42ae-be25-5d0e914fbb49\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.876840 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fh687\" (UniqueName: \"kubernetes.io/projected/e0661780-ca2b-42ae-be25-5d0e914fbb49-kube-api-access-fh687\") pod \"e0661780-ca2b-42ae-be25-5d0e914fbb49\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.876868 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-httpd-config\") pod \"e0661780-ca2b-42ae-be25-5d0e914fbb49\" (UID: \"e0661780-ca2b-42ae-be25-5d0e914fbb49\") " Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.884075 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0661780-ca2b-42ae-be25-5d0e914fbb49-kube-api-access-fh687" (OuterVolumeSpecName: "kube-api-access-fh687") pod "e0661780-ca2b-42ae-be25-5d0e914fbb49" (UID: "e0661780-ca2b-42ae-be25-5d0e914fbb49"). InnerVolumeSpecName "kube-api-access-fh687". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.929115 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "e0661780-ca2b-42ae-be25-5d0e914fbb49" (UID: "e0661780-ca2b-42ae-be25-5d0e914fbb49"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.942418 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e0661780-ca2b-42ae-be25-5d0e914fbb49" (UID: "e0661780-ca2b-42ae-be25-5d0e914fbb49"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.966054 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0661780-ca2b-42ae-be25-5d0e914fbb49" (UID: "e0661780-ca2b-42ae-be25-5d0e914fbb49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.978463 4948 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.978491 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.978501 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fh687\" (UniqueName: \"kubernetes.io/projected/e0661780-ca2b-42ae-be25-5d0e914fbb49-kube-api-access-fh687\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.978512 4948 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.991235 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-config" (OuterVolumeSpecName: "config") pod "e0661780-ca2b-42ae-be25-5d0e914fbb49" (UID: "e0661780-ca2b-42ae-be25-5d0e914fbb49"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:15 crc kubenswrapper[4948]: I0220 08:23:15.994937 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "e0661780-ca2b-42ae-be25-5d0e914fbb49" (UID: "e0661780-ca2b-42ae-be25-5d0e914fbb49"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.005122 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e0661780-ca2b-42ae-be25-5d0e914fbb49" (UID: "e0661780-ca2b-42ae-be25-5d0e914fbb49"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.079863 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.080095 4948 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.080183 4948 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0661780-ca2b-42ae-be25-5d0e914fbb49-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.416798 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerStarted","Data":"e85736cf9a602b496a8372a3fe185272b09bf09e11882adec2e63f62c62bf3d9"} Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.416961 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.421098 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685bb7f5c-jd9nm" event={"ID":"e0661780-ca2b-42ae-be25-5d0e914fbb49","Type":"ContainerDied","Data":"4885fc2e4d8fe9fc0c29512f6a5fd4f1b7a87efbf9c10fe17b8bf4fe7f1ca02a"} Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.421173 4948 scope.go:117] "RemoveContainer" containerID="d2ca1451e75be7daafc6907a8305baa229ed6441f1f79895bd2df00e5c17bfa7" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.421235 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerName="cinder-scheduler" containerID="cri-o://0c8a24ef5e8c2ed45ddfc6908331dbda28dd167039cb8bc7f442c18afef4d964" gracePeriod=30 Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.421259 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-685bb7f5c-jd9nm" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.421272 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerName="probe" containerID="cri-o://c1fbafc4d60f4668a3ce40442dbec31bba82bcb8754ebb88a8aed726953af8e7" gracePeriod=30 Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.452945 4948 scope.go:117] "RemoveContainer" containerID="1ad25f5bfef4b0fa06a939cc9fb8e5bfa3dad488d3265a3cd5fa655c097dda5b" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.457822 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.714538234 podStartE2EDuration="5.457802438s" podCreationTimestamp="2026-02-20 08:23:11 +0000 UTC" firstStartedPulling="2026-02-20 08:23:12.465250595 +0000 UTC m=+1041.439745415" lastFinishedPulling="2026-02-20 08:23:16.208514799 +0000 UTC m=+1045.183009619" observedRunningTime="2026-02-20 08:23:16.450094608 +0000 UTC m=+1045.424589438" watchObservedRunningTime="2026-02-20 08:23:16.457802438 +0000 UTC m=+1045.432297268" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.473618 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-685bb7f5c-jd9nm"] Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.481095 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-685bb7f5c-jd9nm"] Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.693753 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.769187 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-v8n5p"] Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.770127 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" podUID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerName="dnsmasq-dns" containerID="cri-o://ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386" gracePeriod=10 Feb 20 08:23:16 crc kubenswrapper[4948]: I0220 08:23:16.811550 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" podUID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.154:5353: connect: connection refused" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.418926 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.433318 4948 generic.go:334] "Generic (PLEG): container finished" podID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerID="c1fbafc4d60f4668a3ce40442dbec31bba82bcb8754ebb88a8aed726953af8e7" exitCode=0 Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.433393 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"75294aee-c7f3-4b0b-82fc-df058490b90f","Type":"ContainerDied","Data":"c1fbafc4d60f4668a3ce40442dbec31bba82bcb8754ebb88a8aed726953af8e7"} Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.439874 4948 generic.go:334] "Generic (PLEG): container finished" podID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerID="ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386" exitCode=0 Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.439993 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.439996 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" event={"ID":"0cc4ba01-323f-4bcb-9bc3-d639514f6088","Type":"ContainerDied","Data":"ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386"} Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.440152 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-v8n5p" event={"ID":"0cc4ba01-323f-4bcb-9bc3-d639514f6088","Type":"ContainerDied","Data":"5ba21a5e281f6be663a46ae25fc6e9402ee9bcd0e6caec5c0c3b4658222b76c1"} Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.440188 4948 scope.go:117] "RemoveContainer" containerID="ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.470891 4948 scope.go:117] "RemoveContainer" containerID="7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.502324 4948 scope.go:117] "RemoveContainer" containerID="ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386" Feb 20 08:23:17 crc kubenswrapper[4948]: E0220 08:23:17.502715 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386\": container with ID starting with ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386 not found: ID does not exist" containerID="ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.502753 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386"} err="failed to get container status \"ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386\": rpc error: code = NotFound desc = could not find container \"ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386\": container with ID starting with ce123808c7ffa9e9792ff9cd3fa77559a772ebe4ee4f08ad46bdfe3f21ea2386 not found: ID does not exist" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.502776 4948 scope.go:117] "RemoveContainer" containerID="7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743" Feb 20 08:23:17 crc kubenswrapper[4948]: E0220 08:23:17.503199 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743\": container with ID starting with 7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743 not found: ID does not exist" containerID="7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.503235 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743"} err="failed to get container status \"7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743\": rpc error: code = NotFound desc = could not find container \"7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743\": container with ID starting with 7121b5f51a9b7787780e4d13a76cf2b72d32878e2e19ab652f2bed18d9c68743 not found: ID does not exist" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.509150 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-swift-storage-0\") pod \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.509191 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgbxd\" (UniqueName: \"kubernetes.io/projected/0cc4ba01-323f-4bcb-9bc3-d639514f6088-kube-api-access-sgbxd\") pod \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.509246 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-svc\") pod \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.509276 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-config\") pod \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.509488 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-sb\") pod \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.509515 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-nb\") pod \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\" (UID: \"0cc4ba01-323f-4bcb-9bc3-d639514f6088\") " Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.516297 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cc4ba01-323f-4bcb-9bc3-d639514f6088-kube-api-access-sgbxd" (OuterVolumeSpecName: "kube-api-access-sgbxd") pod "0cc4ba01-323f-4bcb-9bc3-d639514f6088" (UID: "0cc4ba01-323f-4bcb-9bc3-d639514f6088"). InnerVolumeSpecName "kube-api-access-sgbxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.554408 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0cc4ba01-323f-4bcb-9bc3-d639514f6088" (UID: "0cc4ba01-323f-4bcb-9bc3-d639514f6088"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.559311 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0cc4ba01-323f-4bcb-9bc3-d639514f6088" (UID: "0cc4ba01-323f-4bcb-9bc3-d639514f6088"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.560495 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0cc4ba01-323f-4bcb-9bc3-d639514f6088" (UID: "0cc4ba01-323f-4bcb-9bc3-d639514f6088"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.564240 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0cc4ba01-323f-4bcb-9bc3-d639514f6088" (UID: "0cc4ba01-323f-4bcb-9bc3-d639514f6088"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.577883 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-config" (OuterVolumeSpecName: "config") pod "0cc4ba01-323f-4bcb-9bc3-d639514f6088" (UID: "0cc4ba01-323f-4bcb-9bc3-d639514f6088"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.610956 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.611011 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgbxd\" (UniqueName: \"kubernetes.io/projected/0cc4ba01-323f-4bcb-9bc3-d639514f6088-kube-api-access-sgbxd\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.611025 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.611036 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.611046 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.611055 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cc4ba01-323f-4bcb-9bc3-d639514f6088-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.732528 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" path="/var/lib/kubelet/pods/e0661780-ca2b-42ae-be25-5d0e914fbb49/volumes" Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.774934 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-v8n5p"] Feb 20 08:23:17 crc kubenswrapper[4948]: I0220 08:23:17.782470 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-v8n5p"] Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.454230 4948 generic.go:334] "Generic (PLEG): container finished" podID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerID="0c8a24ef5e8c2ed45ddfc6908331dbda28dd167039cb8bc7f442c18afef4d964" exitCode=0 Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.454310 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"75294aee-c7f3-4b0b-82fc-df058490b90f","Type":"ContainerDied","Data":"0c8a24ef5e8c2ed45ddfc6908331dbda28dd167039cb8bc7f442c18afef4d964"} Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.828263 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.838607 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vllr\" (UniqueName: \"kubernetes.io/projected/75294aee-c7f3-4b0b-82fc-df058490b90f-kube-api-access-8vllr\") pod \"75294aee-c7f3-4b0b-82fc-df058490b90f\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.838672 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data-custom\") pod \"75294aee-c7f3-4b0b-82fc-df058490b90f\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.838705 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-combined-ca-bundle\") pod \"75294aee-c7f3-4b0b-82fc-df058490b90f\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.838724 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data\") pod \"75294aee-c7f3-4b0b-82fc-df058490b90f\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.838762 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-scripts\") pod \"75294aee-c7f3-4b0b-82fc-df058490b90f\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.838780 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75294aee-c7f3-4b0b-82fc-df058490b90f-etc-machine-id\") pod \"75294aee-c7f3-4b0b-82fc-df058490b90f\" (UID: \"75294aee-c7f3-4b0b-82fc-df058490b90f\") " Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.839076 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75294aee-c7f3-4b0b-82fc-df058490b90f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "75294aee-c7f3-4b0b-82fc-df058490b90f" (UID: "75294aee-c7f3-4b0b-82fc-df058490b90f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.847426 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75294aee-c7f3-4b0b-82fc-df058490b90f-kube-api-access-8vllr" (OuterVolumeSpecName: "kube-api-access-8vllr") pod "75294aee-c7f3-4b0b-82fc-df058490b90f" (UID: "75294aee-c7f3-4b0b-82fc-df058490b90f"). InnerVolumeSpecName "kube-api-access-8vllr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.907153 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "75294aee-c7f3-4b0b-82fc-df058490b90f" (UID: "75294aee-c7f3-4b0b-82fc-df058490b90f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.911186 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-scripts" (OuterVolumeSpecName: "scripts") pod "75294aee-c7f3-4b0b-82fc-df058490b90f" (UID: "75294aee-c7f3-4b0b-82fc-df058490b90f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.956922 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vllr\" (UniqueName: \"kubernetes.io/projected/75294aee-c7f3-4b0b-82fc-df058490b90f-kube-api-access-8vllr\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.956959 4948 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.957087 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.957100 4948 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75294aee-c7f3-4b0b-82fc-df058490b90f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:18 crc kubenswrapper[4948]: I0220 08:23:18.988201 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75294aee-c7f3-4b0b-82fc-df058490b90f" (UID: "75294aee-c7f3-4b0b-82fc-df058490b90f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.029410 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.062690 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.072581 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data" (OuterVolumeSpecName: "config-data") pod "75294aee-c7f3-4b0b-82fc-df058490b90f" (UID: "75294aee-c7f3-4b0b-82fc-df058490b90f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.164645 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75294aee-c7f3-4b0b-82fc-df058490b90f-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.313526 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.470556 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"75294aee-c7f3-4b0b-82fc-df058490b90f","Type":"ContainerDied","Data":"3927e56027e8aee8bf743a6fd812a761cbedf51a0e339d17550adee912677f5f"} Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.470623 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.470988 4948 scope.go:117] "RemoveContainer" containerID="c1fbafc4d60f4668a3ce40442dbec31bba82bcb8754ebb88a8aed726953af8e7" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.514465 4948 scope.go:117] "RemoveContainer" containerID="0c8a24ef5e8c2ed45ddfc6908331dbda28dd167039cb8bc7f442c18afef4d964" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.519846 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.530098 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.545873 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:19 crc kubenswrapper[4948]: E0220 08:23:19.546448 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerName="init" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546474 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerName="init" Feb 20 08:23:19 crc kubenswrapper[4948]: E0220 08:23:19.546494 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerName="dnsmasq-dns" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546503 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerName="dnsmasq-dns" Feb 20 08:23:19 crc kubenswrapper[4948]: E0220 08:23:19.546522 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-httpd" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546529 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-httpd" Feb 20 08:23:19 crc kubenswrapper[4948]: E0220 08:23:19.546546 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-api" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546557 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-api" Feb 20 08:23:19 crc kubenswrapper[4948]: E0220 08:23:19.546579 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerName="cinder-scheduler" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546586 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerName="cinder-scheduler" Feb 20 08:23:19 crc kubenswrapper[4948]: E0220 08:23:19.546605 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerName="probe" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546612 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerName="probe" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546815 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerName="probe" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546843 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" containerName="cinder-scheduler" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546860 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-api" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546878 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0661780-ca2b-42ae-be25-5d0e914fbb49" containerName="neutron-httpd" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.546892 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" containerName="dnsmasq-dns" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.548288 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.551000 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.555845 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.571985 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.572577 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.572625 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-config-data\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.572677 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8w4r\" (UniqueName: \"kubernetes.io/projected/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-kube-api-access-m8w4r\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.572765 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.572795 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-scripts\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.625069 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-86887846d-pqjcb"] Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.626362 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.638807 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-86887846d-pqjcb"] Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674574 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674640 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-config-data\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674676 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dsfr\" (UniqueName: \"kubernetes.io/projected/5047ff00-845b-459f-ac89-6cbf997f93fb-kube-api-access-4dsfr\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674697 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674717 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-scripts\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674737 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-config-data\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674766 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8w4r\" (UniqueName: \"kubernetes.io/projected/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-kube-api-access-m8w4r\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674780 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-combined-ca-bundle\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674808 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-public-tls-certs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674849 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674867 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-internal-tls-certs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674884 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-scripts\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.674939 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5047ff00-845b-459f-ac89-6cbf997f93fb-logs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.675049 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.683637 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.684390 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-config-data\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.684450 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-scripts\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.684454 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.695449 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8w4r\" (UniqueName: \"kubernetes.io/projected/2f8a0f7b-ee05-4e93-a591-7486fecf8cd0-kube-api-access-m8w4r\") pod \"cinder-scheduler-0\" (UID: \"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0\") " pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.730952 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cc4ba01-323f-4bcb-9bc3-d639514f6088" path="/var/lib/kubelet/pods/0cc4ba01-323f-4bcb-9bc3-d639514f6088/volumes" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.731684 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75294aee-c7f3-4b0b-82fc-df058490b90f" path="/var/lib/kubelet/pods/75294aee-c7f3-4b0b-82fc-df058490b90f/volumes" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.776321 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-combined-ca-bundle\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.776605 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-public-tls-certs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.776716 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-internal-tls-certs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.776826 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5047ff00-845b-459f-ac89-6cbf997f93fb-logs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.776958 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-config-data\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.777083 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dsfr\" (UniqueName: \"kubernetes.io/projected/5047ff00-845b-459f-ac89-6cbf997f93fb-kube-api-access-4dsfr\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.777157 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-scripts\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.777750 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5047ff00-845b-459f-ac89-6cbf997f93fb-logs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.779944 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-combined-ca-bundle\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.780820 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-internal-tls-certs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.782878 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-scripts\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.785273 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-public-tls-certs\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.794253 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5047ff00-845b-459f-ac89-6cbf997f93fb-config-data\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.798108 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dsfr\" (UniqueName: \"kubernetes.io/projected/5047ff00-845b-459f-ac89-6cbf997f93fb-kube-api-access-4dsfr\") pod \"placement-86887846d-pqjcb\" (UID: \"5047ff00-845b-459f-ac89-6cbf997f93fb\") " pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.873297 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 08:23:19 crc kubenswrapper[4948]: I0220 08:23:19.949428 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:20 crc kubenswrapper[4948]: I0220 08:23:20.442263 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 08:23:20 crc kubenswrapper[4948]: I0220 08:23:20.499067 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0","Type":"ContainerStarted","Data":"7b9175ec0f8b01106dc5de2e15b6e45b6d8ffb54de656411a7dfe063ac9da99b"} Feb 20 08:23:20 crc kubenswrapper[4948]: I0220 08:23:20.524872 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-86887846d-pqjcb"] Feb 20 08:23:21 crc kubenswrapper[4948]: I0220 08:23:21.517579 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0","Type":"ContainerStarted","Data":"9b3fe039662e55fa4b6cee25944a613b3149ab59c484a7308119f066ce7b1370"} Feb 20 08:23:21 crc kubenswrapper[4948]: I0220 08:23:21.520574 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-86887846d-pqjcb" event={"ID":"5047ff00-845b-459f-ac89-6cbf997f93fb","Type":"ContainerStarted","Data":"8d22e37d5f9b6a5c86cc676fc0e1a18b6aef0379685fd530b8da7ae8302c562a"} Feb 20 08:23:21 crc kubenswrapper[4948]: I0220 08:23:21.520602 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-86887846d-pqjcb" event={"ID":"5047ff00-845b-459f-ac89-6cbf997f93fb","Type":"ContainerStarted","Data":"4d24ae4e85b277b60efc36629974351db737751a376318ecdaabb0c430d33103"} Feb 20 08:23:21 crc kubenswrapper[4948]: I0220 08:23:21.520611 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-86887846d-pqjcb" event={"ID":"5047ff00-845b-459f-ac89-6cbf997f93fb","Type":"ContainerStarted","Data":"d25eb689b81e5ffd366d099059c81fcabd95ba9c317c99a3b01984365a662368"} Feb 20 08:23:21 crc kubenswrapper[4948]: I0220 08:23:21.520824 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:21 crc kubenswrapper[4948]: I0220 08:23:21.520869 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:21 crc kubenswrapper[4948]: I0220 08:23:21.548462 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-86887846d-pqjcb" podStartSLOduration=2.548445226 podStartE2EDuration="2.548445226s" podCreationTimestamp="2026-02-20 08:23:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:21.540704465 +0000 UTC m=+1050.515199285" watchObservedRunningTime="2026-02-20 08:23:21.548445226 +0000 UTC m=+1050.522940046" Feb 20 08:23:22 crc kubenswrapper[4948]: I0220 08:23:22.528877 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2f8a0f7b-ee05-4e93-a591-7486fecf8cd0","Type":"ContainerStarted","Data":"f35ed6518f2bf61623327f1157db4180ec50f47b3db49dac1d2fbf7c882e31df"} Feb 20 08:23:22 crc kubenswrapper[4948]: I0220 08:23:22.558108 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.558093732 podStartE2EDuration="3.558093732s" podCreationTimestamp="2026-02-20 08:23:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:22.555861157 +0000 UTC m=+1051.530355977" watchObservedRunningTime="2026-02-20 08:23:22.558093732 +0000 UTC m=+1051.532588552" Feb 20 08:23:22 crc kubenswrapper[4948]: I0220 08:23:22.890483 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6d4b8df464-wspg2" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.500845 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.502130 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.503574 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-4bss6" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.504573 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.517862 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.539510 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.548918 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq46g\" (UniqueName: \"kubernetes.io/projected/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-kube-api-access-sq46g\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.549184 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-openstack-config\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.549280 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.549366 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-openstack-config-secret\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.651723 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-openstack-config\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.651780 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.651809 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-openstack-config-secret\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.652605 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-openstack-config\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.653084 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq46g\" (UniqueName: \"kubernetes.io/projected/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-kube-api-access-sq46g\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.666060 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.666089 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-openstack-config-secret\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.696072 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq46g\" (UniqueName: \"kubernetes.io/projected/b2a76f68-3b4b-4993-aa4f-3b31b4ee1008-kube-api-access-sq46g\") pod \"openstackclient\" (UID: \"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008\") " pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.857156 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 20 08:23:23 crc kubenswrapper[4948]: I0220 08:23:23.898027 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.204576 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.205682 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-686d666564-8tnjb" Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.299151 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-65487cc4d6-7gdvr"] Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.300321 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-65487cc4d6-7gdvr" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api" containerID="cri-o://ab05c86cbd431ad2cc4a3cef37a6fc481837d1b6073c9a80ed0ee92d983240c8" gracePeriod=30 Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.299935 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-65487cc4d6-7gdvr" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api-log" containerID="cri-o://86e858c136de80a807f997b8991e918a47db690a2c7b9fcc3401caa707f989dd" gracePeriod=30 Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.450195 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 20 08:23:24 crc kubenswrapper[4948]: W0220 08:23:24.466296 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2a76f68_3b4b_4993_aa4f_3b31b4ee1008.slice/crio-2a29f53c784327d73782ce53c1e38b1da346ef1d3769ab76fd235912c07e2a35 WatchSource:0}: Error finding container 2a29f53c784327d73782ce53c1e38b1da346ef1d3769ab76fd235912c07e2a35: Status 404 returned error can't find the container with id 2a29f53c784327d73782ce53c1e38b1da346ef1d3769ab76fd235912c07e2a35 Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.558542 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008","Type":"ContainerStarted","Data":"2a29f53c784327d73782ce53c1e38b1da346ef1d3769ab76fd235912c07e2a35"} Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.560582 4948 generic.go:334] "Generic (PLEG): container finished" podID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerID="86e858c136de80a807f997b8991e918a47db690a2c7b9fcc3401caa707f989dd" exitCode=143 Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.560633 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65487cc4d6-7gdvr" event={"ID":"6a669dd0-d287-4d07-9eec-aea5ed80c791","Type":"ContainerDied","Data":"86e858c136de80a807f997b8991e918a47db690a2c7b9fcc3401caa707f989dd"} Feb 20 08:23:24 crc kubenswrapper[4948]: I0220 08:23:24.873961 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 20 08:23:27 crc kubenswrapper[4948]: I0220 08:23:27.466601 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-65487cc4d6-7gdvr" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": read tcp 10.217.0.2:48370->10.217.0.166:9311: read: connection reset by peer" Feb 20 08:23:27 crc kubenswrapper[4948]: I0220 08:23:27.466679 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-65487cc4d6-7gdvr" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.166:9311/healthcheck\": read tcp 10.217.0.2:48382->10.217.0.166:9311: read: connection reset by peer" Feb 20 08:23:27 crc kubenswrapper[4948]: I0220 08:23:27.591152 4948 generic.go:334] "Generic (PLEG): container finished" podID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerID="ab05c86cbd431ad2cc4a3cef37a6fc481837d1b6073c9a80ed0ee92d983240c8" exitCode=0 Feb 20 08:23:27 crc kubenswrapper[4948]: I0220 08:23:27.591364 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65487cc4d6-7gdvr" event={"ID":"6a669dd0-d287-4d07-9eec-aea5ed80c791","Type":"ContainerDied","Data":"ab05c86cbd431ad2cc4a3cef37a6fc481837d1b6073c9a80ed0ee92d983240c8"} Feb 20 08:23:27 crc kubenswrapper[4948]: I0220 08:23:27.947945 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.052082 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-combined-ca-bundle\") pod \"6a669dd0-d287-4d07-9eec-aea5ed80c791\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.052210 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data\") pod \"6a669dd0-d287-4d07-9eec-aea5ed80c791\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.052280 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data-custom\") pod \"6a669dd0-d287-4d07-9eec-aea5ed80c791\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.052379 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld59l\" (UniqueName: \"kubernetes.io/projected/6a669dd0-d287-4d07-9eec-aea5ed80c791-kube-api-access-ld59l\") pod \"6a669dd0-d287-4d07-9eec-aea5ed80c791\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.052423 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a669dd0-d287-4d07-9eec-aea5ed80c791-logs\") pod \"6a669dd0-d287-4d07-9eec-aea5ed80c791\" (UID: \"6a669dd0-d287-4d07-9eec-aea5ed80c791\") " Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.053144 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a669dd0-d287-4d07-9eec-aea5ed80c791-logs" (OuterVolumeSpecName: "logs") pod "6a669dd0-d287-4d07-9eec-aea5ed80c791" (UID: "6a669dd0-d287-4d07-9eec-aea5ed80c791"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.057624 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a669dd0-d287-4d07-9eec-aea5ed80c791-kube-api-access-ld59l" (OuterVolumeSpecName: "kube-api-access-ld59l") pod "6a669dd0-d287-4d07-9eec-aea5ed80c791" (UID: "6a669dd0-d287-4d07-9eec-aea5ed80c791"). InnerVolumeSpecName "kube-api-access-ld59l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.089875 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6a669dd0-d287-4d07-9eec-aea5ed80c791" (UID: "6a669dd0-d287-4d07-9eec-aea5ed80c791"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.110449 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a669dd0-d287-4d07-9eec-aea5ed80c791" (UID: "6a669dd0-d287-4d07-9eec-aea5ed80c791"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.129182 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data" (OuterVolumeSpecName: "config-data") pod "6a669dd0-d287-4d07-9eec-aea5ed80c791" (UID: "6a669dd0-d287-4d07-9eec-aea5ed80c791"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.154245 4948 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.154274 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld59l\" (UniqueName: \"kubernetes.io/projected/6a669dd0-d287-4d07-9eec-aea5ed80c791-kube-api-access-ld59l\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.154286 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a669dd0-d287-4d07-9eec-aea5ed80c791-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.154296 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.154305 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a669dd0-d287-4d07-9eec-aea5ed80c791-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.272184 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.272492 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="proxy-httpd" containerID="cri-o://e85736cf9a602b496a8372a3fe185272b09bf09e11882adec2e63f62c62bf3d9" gracePeriod=30 Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.272584 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="sg-core" containerID="cri-o://d829cc037c7ba0bd0abfe863863495a3388592344985fae656879df8c40ec990" gracePeriod=30 Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.272789 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="ceilometer-notification-agent" containerID="cri-o://7e6d02d6234c2e72678f463fac466a22ef0548ef1c161d0d356079f01463ff09" gracePeriod=30 Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.272859 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="ceilometer-central-agent" containerID="cri-o://eab93963947d8aecd2b45387be911464a6576e8236c9ccd3d0a009ad65054a74" gracePeriod=30 Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.289287 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.167:3000/\": EOF" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.412126 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-x2pm7"] Feb 20 08:23:28 crc kubenswrapper[4948]: E0220 08:23:28.412847 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api-log" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.412868 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api-log" Feb 20 08:23:28 crc kubenswrapper[4948]: E0220 08:23:28.412884 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.412891 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.413083 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.413102 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" containerName="barbican-api-log" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.413658 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.434544 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-x2pm7"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.460709 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa398740-9a54-472f-9736-0959dd54b657-operator-scripts\") pod \"nova-api-db-create-x2pm7\" (UID: \"fa398740-9a54-472f-9736-0959dd54b657\") " pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.460777 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rx99\" (UniqueName: \"kubernetes.io/projected/fa398740-9a54-472f-9736-0959dd54b657-kube-api-access-7rx99\") pod \"nova-api-db-create-x2pm7\" (UID: \"fa398740-9a54-472f-9736-0959dd54b657\") " pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.562552 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rx99\" (UniqueName: \"kubernetes.io/projected/fa398740-9a54-472f-9736-0959dd54b657-kube-api-access-7rx99\") pod \"nova-api-db-create-x2pm7\" (UID: \"fa398740-9a54-472f-9736-0959dd54b657\") " pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.562761 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa398740-9a54-472f-9736-0959dd54b657-operator-scripts\") pod \"nova-api-db-create-x2pm7\" (UID: \"fa398740-9a54-472f-9736-0959dd54b657\") " pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.563808 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa398740-9a54-472f-9736-0959dd54b657-operator-scripts\") pod \"nova-api-db-create-x2pm7\" (UID: \"fa398740-9a54-472f-9736-0959dd54b657\") " pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.592836 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rx99\" (UniqueName: \"kubernetes.io/projected/fa398740-9a54-472f-9736-0959dd54b657-kube-api-access-7rx99\") pod \"nova-api-db-create-x2pm7\" (UID: \"fa398740-9a54-472f-9736-0959dd54b657\") " pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.612829 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65487cc4d6-7gdvr" event={"ID":"6a669dd0-d287-4d07-9eec-aea5ed80c791","Type":"ContainerDied","Data":"b3b840325250a810c158ca1f61d60db618feec352dbf68e16aad14e15139327d"} Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.612888 4948 scope.go:117] "RemoveContainer" containerID="ab05c86cbd431ad2cc4a3cef37a6fc481837d1b6073c9a80ed0ee92d983240c8" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.613030 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65487cc4d6-7gdvr" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.630532 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7c7cd9876c-262qf"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.635311 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.642473 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.642732 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.643406 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.644153 4948 generic.go:334] "Generic (PLEG): container finished" podID="44092649-be82-4225-8cfc-d37dbe840877" containerID="e85736cf9a602b496a8372a3fe185272b09bf09e11882adec2e63f62c62bf3d9" exitCode=0 Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.644172 4948 generic.go:334] "Generic (PLEG): container finished" podID="44092649-be82-4225-8cfc-d37dbe840877" containerID="d829cc037c7ba0bd0abfe863863495a3388592344985fae656879df8c40ec990" exitCode=2 Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.644190 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerDied","Data":"e85736cf9a602b496a8372a3fe185272b09bf09e11882adec2e63f62c62bf3d9"} Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.644212 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerDied","Data":"d829cc037c7ba0bd0abfe863863495a3388592344985fae656879df8c40ec990"} Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.664161 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-config-data\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.664219 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-combined-ca-bundle\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.664699 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgtbs\" (UniqueName: \"kubernetes.io/projected/30f48966-e44d-44bc-a868-7e8119ac186c-kube-api-access-lgtbs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.664753 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30f48966-e44d-44bc-a868-7e8119ac186c-run-httpd\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.664769 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-public-tls-certs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.664788 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/30f48966-e44d-44bc-a868-7e8119ac186c-etc-swift\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.664835 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-internal-tls-certs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.664850 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30f48966-e44d-44bc-a868-7e8119ac186c-log-httpd\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.670068 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-mgnlb"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.672926 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.701208 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-mgnlb"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.716811 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7c7cd9876c-262qf"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.728204 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-aecf-account-create-update-gnnvg"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.729603 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.731879 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.747945 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-aecf-account-create-update-gnnvg"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.750196 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.755530 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-65487cc4d6-7gdvr"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.765995 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-65487cc4d6-7gdvr"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766202 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-internal-tls-certs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766253 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b61242-b8c1-4e1b-b319-3eafb79ef048-operator-scripts\") pod \"nova-cell0-db-create-mgnlb\" (UID: \"20b61242-b8c1-4e1b-b319-3eafb79ef048\") " pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766283 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30f48966-e44d-44bc-a868-7e8119ac186c-log-httpd\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766366 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-config-data\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766417 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-operator-scripts\") pod \"nova-api-aecf-account-create-update-gnnvg\" (UID: \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\") " pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766457 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-combined-ca-bundle\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766490 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgtbs\" (UniqueName: \"kubernetes.io/projected/30f48966-e44d-44bc-a868-7e8119ac186c-kube-api-access-lgtbs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766554 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30f48966-e44d-44bc-a868-7e8119ac186c-run-httpd\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766579 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-public-tls-certs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766604 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/30f48966-e44d-44bc-a868-7e8119ac186c-etc-swift\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766643 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ssjs\" (UniqueName: \"kubernetes.io/projected/20b61242-b8c1-4e1b-b319-3eafb79ef048-kube-api-access-2ssjs\") pod \"nova-cell0-db-create-mgnlb\" (UID: \"20b61242-b8c1-4e1b-b319-3eafb79ef048\") " pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766679 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ttpc\" (UniqueName: \"kubernetes.io/projected/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-kube-api-access-7ttpc\") pod \"nova-api-aecf-account-create-update-gnnvg\" (UID: \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\") " pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.766838 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30f48966-e44d-44bc-a868-7e8119ac186c-log-httpd\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.767134 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30f48966-e44d-44bc-a868-7e8119ac186c-run-httpd\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.771457 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-internal-tls-certs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.772067 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/30f48966-e44d-44bc-a868-7e8119ac186c-etc-swift\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.783866 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-public-tls-certs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.784381 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-combined-ca-bundle\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.788394 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30f48966-e44d-44bc-a868-7e8119ac186c-config-data\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.788662 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgtbs\" (UniqueName: \"kubernetes.io/projected/30f48966-e44d-44bc-a868-7e8119ac186c-kube-api-access-lgtbs\") pod \"swift-proxy-7c7cd9876c-262qf\" (UID: \"30f48966-e44d-44bc-a868-7e8119ac186c\") " pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.798469 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-2rgc6"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.807620 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.821167 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-2rgc6"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.841490 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-78de-account-create-update-fwjdc"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.842805 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.844418 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.849636 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-78de-account-create-update-fwjdc"] Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.867365 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b61242-b8c1-4e1b-b319-3eafb79ef048-operator-scripts\") pod \"nova-cell0-db-create-mgnlb\" (UID: \"20b61242-b8c1-4e1b-b319-3eafb79ef048\") " pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.867476 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpq6v\" (UniqueName: \"kubernetes.io/projected/8b167b56-655a-4220-9823-8cc606f5f034-kube-api-access-fpq6v\") pod \"nova-cell0-78de-account-create-update-fwjdc\" (UID: \"8b167b56-655a-4220-9823-8cc606f5f034\") " pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.867535 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hgw6\" (UniqueName: \"kubernetes.io/projected/a5cb8c15-68fc-47db-9f68-1a8401f99f88-kube-api-access-2hgw6\") pod \"nova-cell1-db-create-2rgc6\" (UID: \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\") " pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.867556 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-operator-scripts\") pod \"nova-api-aecf-account-create-update-gnnvg\" (UID: \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\") " pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.867609 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5cb8c15-68fc-47db-9f68-1a8401f99f88-operator-scripts\") pod \"nova-cell1-db-create-2rgc6\" (UID: \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\") " pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.867655 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ssjs\" (UniqueName: \"kubernetes.io/projected/20b61242-b8c1-4e1b-b319-3eafb79ef048-kube-api-access-2ssjs\") pod \"nova-cell0-db-create-mgnlb\" (UID: \"20b61242-b8c1-4e1b-b319-3eafb79ef048\") " pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.867681 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b167b56-655a-4220-9823-8cc606f5f034-operator-scripts\") pod \"nova-cell0-78de-account-create-update-fwjdc\" (UID: \"8b167b56-655a-4220-9823-8cc606f5f034\") " pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.867699 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ttpc\" (UniqueName: \"kubernetes.io/projected/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-kube-api-access-7ttpc\") pod \"nova-api-aecf-account-create-update-gnnvg\" (UID: \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\") " pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.868245 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b61242-b8c1-4e1b-b319-3eafb79ef048-operator-scripts\") pod \"nova-cell0-db-create-mgnlb\" (UID: \"20b61242-b8c1-4e1b-b319-3eafb79ef048\") " pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.868584 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-operator-scripts\") pod \"nova-api-aecf-account-create-update-gnnvg\" (UID: \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\") " pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.886963 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ttpc\" (UniqueName: \"kubernetes.io/projected/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-kube-api-access-7ttpc\") pod \"nova-api-aecf-account-create-update-gnnvg\" (UID: \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\") " pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.894728 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ssjs\" (UniqueName: \"kubernetes.io/projected/20b61242-b8c1-4e1b-b319-3eafb79ef048-kube-api-access-2ssjs\") pod \"nova-cell0-db-create-mgnlb\" (UID: \"20b61242-b8c1-4e1b-b319-3eafb79ef048\") " pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.956731 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.970658 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b167b56-655a-4220-9823-8cc606f5f034-operator-scripts\") pod \"nova-cell0-78de-account-create-update-fwjdc\" (UID: \"8b167b56-655a-4220-9823-8cc606f5f034\") " pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.971498 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b167b56-655a-4220-9823-8cc606f5f034-operator-scripts\") pod \"nova-cell0-78de-account-create-update-fwjdc\" (UID: \"8b167b56-655a-4220-9823-8cc606f5f034\") " pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.971944 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpq6v\" (UniqueName: \"kubernetes.io/projected/8b167b56-655a-4220-9823-8cc606f5f034-kube-api-access-fpq6v\") pod \"nova-cell0-78de-account-create-update-fwjdc\" (UID: \"8b167b56-655a-4220-9823-8cc606f5f034\") " pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.972841 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hgw6\" (UniqueName: \"kubernetes.io/projected/a5cb8c15-68fc-47db-9f68-1a8401f99f88-kube-api-access-2hgw6\") pod \"nova-cell1-db-create-2rgc6\" (UID: \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\") " pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.973267 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5cb8c15-68fc-47db-9f68-1a8401f99f88-operator-scripts\") pod \"nova-cell1-db-create-2rgc6\" (UID: \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\") " pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.975287 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5cb8c15-68fc-47db-9f68-1a8401f99f88-operator-scripts\") pod \"nova-cell1-db-create-2rgc6\" (UID: \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\") " pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.988732 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.988951 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpq6v\" (UniqueName: \"kubernetes.io/projected/8b167b56-655a-4220-9823-8cc606f5f034-kube-api-access-fpq6v\") pod \"nova-cell0-78de-account-create-update-fwjdc\" (UID: \"8b167b56-655a-4220-9823-8cc606f5f034\") " pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:28 crc kubenswrapper[4948]: I0220 08:23:28.996454 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hgw6\" (UniqueName: \"kubernetes.io/projected/a5cb8c15-68fc-47db-9f68-1a8401f99f88-kube-api-access-2hgw6\") pod \"nova-cell1-db-create-2rgc6\" (UID: \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\") " pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.028120 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-ca33-account-create-update-dwqpd"] Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.029672 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.031681 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.039070 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ca33-account-create-update-dwqpd"] Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.051621 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.177280 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-operator-scripts\") pod \"nova-cell1-ca33-account-create-update-dwqpd\" (UID: \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\") " pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.177364 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl2t2\" (UniqueName: \"kubernetes.io/projected/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-kube-api-access-vl2t2\") pod \"nova-cell1-ca33-account-create-update-dwqpd\" (UID: \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\") " pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.188277 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.237665 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.279026 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-operator-scripts\") pod \"nova-cell1-ca33-account-create-update-dwqpd\" (UID: \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\") " pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.279119 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl2t2\" (UniqueName: \"kubernetes.io/projected/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-kube-api-access-vl2t2\") pod \"nova-cell1-ca33-account-create-update-dwqpd\" (UID: \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\") " pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.279921 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-operator-scripts\") pod \"nova-cell1-ca33-account-create-update-dwqpd\" (UID: \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\") " pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.297404 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl2t2\" (UniqueName: \"kubernetes.io/projected/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-kube-api-access-vl2t2\") pod \"nova-cell1-ca33-account-create-update-dwqpd\" (UID: \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\") " pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.381928 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.657247 4948 generic.go:334] "Generic (PLEG): container finished" podID="44092649-be82-4225-8cfc-d37dbe840877" containerID="eab93963947d8aecd2b45387be911464a6576e8236c9ccd3d0a009ad65054a74" exitCode=0 Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.657284 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerDied","Data":"eab93963947d8aecd2b45387be911464a6576e8236c9ccd3d0a009ad65054a74"} Feb 20 08:23:29 crc kubenswrapper[4948]: I0220 08:23:29.735961 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a669dd0-d287-4d07-9eec-aea5ed80c791" path="/var/lib/kubelet/pods/6a669dd0-d287-4d07-9eec-aea5ed80c791/volumes" Feb 20 08:23:30 crc kubenswrapper[4948]: I0220 08:23:30.092399 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 20 08:23:31 crc kubenswrapper[4948]: I0220 08:23:31.678460 4948 generic.go:334] "Generic (PLEG): container finished" podID="44092649-be82-4225-8cfc-d37dbe840877" containerID="7e6d02d6234c2e72678f463fac466a22ef0548ef1c161d0d356079f01463ff09" exitCode=0 Feb 20 08:23:31 crc kubenswrapper[4948]: I0220 08:23:31.678535 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerDied","Data":"7e6d02d6234c2e72678f463fac466a22ef0548ef1c161d0d356079f01463ff09"} Feb 20 08:23:33 crc kubenswrapper[4948]: I0220 08:23:33.870656 4948 scope.go:117] "RemoveContainer" containerID="86e858c136de80a807f997b8991e918a47db690a2c7b9fcc3401caa707f989dd" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.327023 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.470355 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-config-data\") pod \"44092649-be82-4225-8cfc-d37dbe840877\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.470402 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-sg-core-conf-yaml\") pod \"44092649-be82-4225-8cfc-d37dbe840877\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.470506 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-scripts\") pod \"44092649-be82-4225-8cfc-d37dbe840877\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.470526 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-combined-ca-bundle\") pod \"44092649-be82-4225-8cfc-d37dbe840877\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.470633 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-log-httpd\") pod \"44092649-be82-4225-8cfc-d37dbe840877\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.470667 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rkd9\" (UniqueName: \"kubernetes.io/projected/44092649-be82-4225-8cfc-d37dbe840877-kube-api-access-4rkd9\") pod \"44092649-be82-4225-8cfc-d37dbe840877\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.470689 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-run-httpd\") pod \"44092649-be82-4225-8cfc-d37dbe840877\" (UID: \"44092649-be82-4225-8cfc-d37dbe840877\") " Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.471511 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "44092649-be82-4225-8cfc-d37dbe840877" (UID: "44092649-be82-4225-8cfc-d37dbe840877"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.472240 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "44092649-be82-4225-8cfc-d37dbe840877" (UID: "44092649-be82-4225-8cfc-d37dbe840877"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.477629 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44092649-be82-4225-8cfc-d37dbe840877-kube-api-access-4rkd9" (OuterVolumeSpecName: "kube-api-access-4rkd9") pod "44092649-be82-4225-8cfc-d37dbe840877" (UID: "44092649-be82-4225-8cfc-d37dbe840877"). InnerVolumeSpecName "kube-api-access-4rkd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.483036 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-scripts" (OuterVolumeSpecName: "scripts") pod "44092649-be82-4225-8cfc-d37dbe840877" (UID: "44092649-be82-4225-8cfc-d37dbe840877"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.506921 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "44092649-be82-4225-8cfc-d37dbe840877" (UID: "44092649-be82-4225-8cfc-d37dbe840877"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.558807 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "44092649-be82-4225-8cfc-d37dbe840877" (UID: "44092649-be82-4225-8cfc-d37dbe840877"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.572886 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rkd9\" (UniqueName: \"kubernetes.io/projected/44092649-be82-4225-8cfc-d37dbe840877-kube-api-access-4rkd9\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.572920 4948 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.572929 4948 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.572941 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.572949 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.572957 4948 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/44092649-be82-4225-8cfc-d37dbe840877-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.603092 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-config-data" (OuterVolumeSpecName: "config-data") pod "44092649-be82-4225-8cfc-d37dbe840877" (UID: "44092649-be82-4225-8cfc-d37dbe840877"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.638309 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-2rgc6"] Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.649023 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-aecf-account-create-update-gnnvg"] Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.674869 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44092649-be82-4225-8cfc-d37dbe840877-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.717869 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b2a76f68-3b4b-4993-aa4f-3b31b4ee1008","Type":"ContainerStarted","Data":"b1aae9b1fd753863bdd28b79973fb865f9bfe71025414314382214f9398365f4"} Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.720391 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2rgc6" event={"ID":"a5cb8c15-68fc-47db-9f68-1a8401f99f88","Type":"ContainerStarted","Data":"7de096242d5ba0d916d6903858d689e5e9254091b9036d8255334ff0ea8f1f16"} Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.726821 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"44092649-be82-4225-8cfc-d37dbe840877","Type":"ContainerDied","Data":"8266937b599da958a185777fc1969453e86491578f18e8d8f35aa2e0681bba26"} Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.726829 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.726878 4948 scope.go:117] "RemoveContainer" containerID="e85736cf9a602b496a8372a3fe185272b09bf09e11882adec2e63f62c62bf3d9" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.730650 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-aecf-account-create-update-gnnvg" event={"ID":"8a311ded-ab56-4f07-8b7f-9dcc4d70f647","Type":"ContainerStarted","Data":"04a106975b06c46c735fa10604e40565f439a58073da7467ca0316db88eec4f9"} Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.740097 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.204848791 podStartE2EDuration="11.740078988s" podCreationTimestamp="2026-02-20 08:23:23 +0000 UTC" firstStartedPulling="2026-02-20 08:23:24.468637252 +0000 UTC m=+1053.443132072" lastFinishedPulling="2026-02-20 08:23:34.003867449 +0000 UTC m=+1062.978362269" observedRunningTime="2026-02-20 08:23:34.733116166 +0000 UTC m=+1063.707611006" watchObservedRunningTime="2026-02-20 08:23:34.740078988 +0000 UTC m=+1063.714573808" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.778152 4948 scope.go:117] "RemoveContainer" containerID="d829cc037c7ba0bd0abfe863863495a3388592344985fae656879df8c40ec990" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.779525 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-mgnlb"] Feb 20 08:23:34 crc kubenswrapper[4948]: W0220 08:23:34.781985 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20b61242_b8c1_4e1b_b319_3eafb79ef048.slice/crio-c5c2d7e3e5610297feff1dafc766462b7a8ac2739d1c5be80ff63c9dbf1bcc12 WatchSource:0}: Error finding container c5c2d7e3e5610297feff1dafc766462b7a8ac2739d1c5be80ff63c9dbf1bcc12: Status 404 returned error can't find the container with id c5c2d7e3e5610297feff1dafc766462b7a8ac2739d1c5be80ff63c9dbf1bcc12 Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.832098 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.851798 4948 scope.go:117] "RemoveContainer" containerID="7e6d02d6234c2e72678f463fac466a22ef0548ef1c161d0d356079f01463ff09" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.855589 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.876399 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:34 crc kubenswrapper[4948]: E0220 08:23:34.876840 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="ceilometer-central-agent" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.876856 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="ceilometer-central-agent" Feb 20 08:23:34 crc kubenswrapper[4948]: E0220 08:23:34.876870 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="sg-core" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.876878 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="sg-core" Feb 20 08:23:34 crc kubenswrapper[4948]: E0220 08:23:34.876908 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="proxy-httpd" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.876916 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="proxy-httpd" Feb 20 08:23:34 crc kubenswrapper[4948]: E0220 08:23:34.876937 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="ceilometer-notification-agent" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.876944 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="ceilometer-notification-agent" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.877163 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="ceilometer-central-agent" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.877185 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="sg-core" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.877193 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="proxy-httpd" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.877215 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="44092649-be82-4225-8cfc-d37dbe840877" containerName="ceilometer-notification-agent" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.883695 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.885271 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.888341 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.888624 4948 scope.go:117] "RemoveContainer" containerID="eab93963947d8aecd2b45387be911464a6576e8236c9ccd3d0a009ad65054a74" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.888908 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.892989 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-78de-account-create-update-fwjdc"] Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.919168 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7c7cd9876c-262qf"] Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.929846 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-x2pm7"] Feb 20 08:23:34 crc kubenswrapper[4948]: I0220 08:23:34.945999 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ca33-account-create-update-dwqpd"] Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.081082 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.081184 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-run-httpd\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.081216 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-config-data\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.081400 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-scripts\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.081516 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.081626 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc4df\" (UniqueName: \"kubernetes.io/projected/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-kube-api-access-qc4df\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.081681 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-log-httpd\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.182869 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.183049 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-run-httpd\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.183089 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-config-data\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.183128 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-scripts\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.183167 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.183194 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc4df\" (UniqueName: \"kubernetes.io/projected/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-kube-api-access-qc4df\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.183232 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-log-httpd\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.183890 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-log-httpd\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.184013 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-run-httpd\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.188060 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-config-data\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.188886 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.189537 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-scripts\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.193556 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.202562 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc4df\" (UniqueName: \"kubernetes.io/projected/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-kube-api-access-qc4df\") pod \"ceilometer-0\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.363683 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.735106 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44092649-be82-4225-8cfc-d37dbe840877" path="/var/lib/kubelet/pods/44092649-be82-4225-8cfc-d37dbe840877/volumes" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.749878 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" event={"ID":"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f","Type":"ContainerStarted","Data":"97e6bc167799c61927aab7b946388571d89bdf96214e87d2ba307142c0a90dd5"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.749926 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" event={"ID":"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f","Type":"ContainerStarted","Data":"4a6359d724bfd5ae6bb40a902ffdabe96c01743e8c81b47b0e26ad235b7eaeff"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.753842 4948 generic.go:334] "Generic (PLEG): container finished" podID="8a311ded-ab56-4f07-8b7f-9dcc4d70f647" containerID="dd9bca0f8ae0d62df16de42592e7d1d0793a43cd05a5a5d5dcaf73199e707005" exitCode=0 Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.754218 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-aecf-account-create-update-gnnvg" event={"ID":"8a311ded-ab56-4f07-8b7f-9dcc4d70f647","Type":"ContainerDied","Data":"dd9bca0f8ae0d62df16de42592e7d1d0793a43cd05a5a5d5dcaf73199e707005"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.766317 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x2pm7" event={"ID":"fa398740-9a54-472f-9736-0959dd54b657","Type":"ContainerStarted","Data":"be6725bc5e2da748cc92ca6d3d54cea0585e5ca878e37159db354dc55a1eae54"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.766521 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x2pm7" event={"ID":"fa398740-9a54-472f-9736-0959dd54b657","Type":"ContainerStarted","Data":"372eccf775e290867aea35db34cd53f0b147aa6e61ff520ef1ac676e72e0326c"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.782137 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" podStartSLOduration=6.782116785 podStartE2EDuration="6.782116785s" podCreationTimestamp="2026-02-20 08:23:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:35.76371817 +0000 UTC m=+1064.738212990" watchObservedRunningTime="2026-02-20 08:23:35.782116785 +0000 UTC m=+1064.756611605" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.784845 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7c7cd9876c-262qf" event={"ID":"30f48966-e44d-44bc-a868-7e8119ac186c","Type":"ContainerStarted","Data":"811eeb0baa418ab2012f0916ae52adf1815521bd676db6d446531d2de318d57e"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.784992 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7c7cd9876c-262qf" event={"ID":"30f48966-e44d-44bc-a868-7e8119ac186c","Type":"ContainerStarted","Data":"5bc8f894cd28a290b6306bff2fa7f5125cf755eb2a8fdd1a4c96b9198eadef20"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.785071 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7c7cd9876c-262qf" event={"ID":"30f48966-e44d-44bc-a868-7e8119ac186c","Type":"ContainerStarted","Data":"5f5a0fa2e5552aae0ce9199499829b8bae2123b19a32b32e0c4140636cd5b2cf"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.785884 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.785986 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.802076 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-x2pm7" podStartSLOduration=7.802059238 podStartE2EDuration="7.802059238s" podCreationTimestamp="2026-02-20 08:23:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:35.801674869 +0000 UTC m=+1064.776169689" watchObservedRunningTime="2026-02-20 08:23:35.802059238 +0000 UTC m=+1064.776554058" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.806069 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-mgnlb" event={"ID":"20b61242-b8c1-4e1b-b319-3eafb79ef048","Type":"ContainerStarted","Data":"ca086fdde2d4c69122e9c700d4d4af3b8218c685639bf6df2d417cc08c1e51d3"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.806109 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-mgnlb" event={"ID":"20b61242-b8c1-4e1b-b319-3eafb79ef048","Type":"ContainerStarted","Data":"c5c2d7e3e5610297feff1dafc766462b7a8ac2739d1c5be80ff63c9dbf1bcc12"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.810395 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" event={"ID":"8b167b56-655a-4220-9823-8cc606f5f034","Type":"ContainerStarted","Data":"57a582df620782031dd517ee8549cae414c4157e56c0fc270434b874a8cbfef6"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.810454 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" event={"ID":"8b167b56-655a-4220-9823-8cc606f5f034","Type":"ContainerStarted","Data":"c9774d1c3be31829d1732d7995a7c4a40d5cd961beaca133cbe7c41c323af95a"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.832257 4948 generic.go:334] "Generic (PLEG): container finished" podID="a5cb8c15-68fc-47db-9f68-1a8401f99f88" containerID="75747717ce5381d273c3b7d9687d0d98efbbf16a7bbea02c0a10888c6dd3a064" exitCode=0 Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.832367 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2rgc6" event={"ID":"a5cb8c15-68fc-47db-9f68-1a8401f99f88","Type":"ContainerDied","Data":"75747717ce5381d273c3b7d9687d0d98efbbf16a7bbea02c0a10888c6dd3a064"} Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.835481 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.837752 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7c7cd9876c-262qf" podStartSLOduration=7.837739261 podStartE2EDuration="7.837739261s" podCreationTimestamp="2026-02-20 08:23:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:35.822445803 +0000 UTC m=+1064.796940623" watchObservedRunningTime="2026-02-20 08:23:35.837739261 +0000 UTC m=+1064.812234081" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.852639 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-mgnlb" podStartSLOduration=7.85262262 podStartE2EDuration="7.85262262s" podCreationTimestamp="2026-02-20 08:23:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:35.85221861 +0000 UTC m=+1064.826713430" watchObservedRunningTime="2026-02-20 08:23:35.85262262 +0000 UTC m=+1064.827117440" Feb 20 08:23:35 crc kubenswrapper[4948]: I0220 08:23:35.901923 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" podStartSLOduration=7.901898659 podStartE2EDuration="7.901898659s" podCreationTimestamp="2026-02-20 08:23:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:23:35.871281071 +0000 UTC m=+1064.845775891" watchObservedRunningTime="2026-02-20 08:23:35.901898659 +0000 UTC m=+1064.876393479" Feb 20 08:23:35 crc kubenswrapper[4948]: W0220 08:23:35.924782 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeae40be2_f936_4e13_bc8e_0f84bbcd8edd.slice/crio-8b2a81f0ae61bef074923309a4e18638192928b3bbadabe1634088ddab7ac96c WatchSource:0}: Error finding container 8b2a81f0ae61bef074923309a4e18638192928b3bbadabe1634088ddab7ac96c: Status 404 returned error can't find the container with id 8b2a81f0ae61bef074923309a4e18638192928b3bbadabe1634088ddab7ac96c Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.843276 4948 generic.go:334] "Generic (PLEG): container finished" podID="8b167b56-655a-4220-9823-8cc606f5f034" containerID="57a582df620782031dd517ee8549cae414c4157e56c0fc270434b874a8cbfef6" exitCode=0 Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.843375 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" event={"ID":"8b167b56-655a-4220-9823-8cc606f5f034","Type":"ContainerDied","Data":"57a582df620782031dd517ee8549cae414c4157e56c0fc270434b874a8cbfef6"} Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.845748 4948 generic.go:334] "Generic (PLEG): container finished" podID="b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f" containerID="97e6bc167799c61927aab7b946388571d89bdf96214e87d2ba307142c0a90dd5" exitCode=0 Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.845814 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" event={"ID":"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f","Type":"ContainerDied","Data":"97e6bc167799c61927aab7b946388571d89bdf96214e87d2ba307142c0a90dd5"} Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.848373 4948 generic.go:334] "Generic (PLEG): container finished" podID="fa398740-9a54-472f-9736-0959dd54b657" containerID="be6725bc5e2da748cc92ca6d3d54cea0585e5ca878e37159db354dc55a1eae54" exitCode=0 Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.848444 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x2pm7" event={"ID":"fa398740-9a54-472f-9736-0959dd54b657","Type":"ContainerDied","Data":"be6725bc5e2da748cc92ca6d3d54cea0585e5ca878e37159db354dc55a1eae54"} Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.850416 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerStarted","Data":"ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b"} Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.850443 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerStarted","Data":"8b2a81f0ae61bef074923309a4e18638192928b3bbadabe1634088ddab7ac96c"} Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.854419 4948 generic.go:334] "Generic (PLEG): container finished" podID="20b61242-b8c1-4e1b-b319-3eafb79ef048" containerID="ca086fdde2d4c69122e9c700d4d4af3b8218c685639bf6df2d417cc08c1e51d3" exitCode=0 Feb 20 08:23:36 crc kubenswrapper[4948]: I0220 08:23:36.854457 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-mgnlb" event={"ID":"20b61242-b8c1-4e1b-b319-3eafb79ef048","Type":"ContainerDied","Data":"ca086fdde2d4c69122e9c700d4d4af3b8218c685639bf6df2d417cc08c1e51d3"} Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.415446 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.423043 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.524076 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ttpc\" (UniqueName: \"kubernetes.io/projected/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-kube-api-access-7ttpc\") pod \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\" (UID: \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\") " Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.524181 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hgw6\" (UniqueName: \"kubernetes.io/projected/a5cb8c15-68fc-47db-9f68-1a8401f99f88-kube-api-access-2hgw6\") pod \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\" (UID: \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\") " Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.524283 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5cb8c15-68fc-47db-9f68-1a8401f99f88-operator-scripts\") pod \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\" (UID: \"a5cb8c15-68fc-47db-9f68-1a8401f99f88\") " Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.524301 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-operator-scripts\") pod \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\" (UID: \"8a311ded-ab56-4f07-8b7f-9dcc4d70f647\") " Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.525097 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8a311ded-ab56-4f07-8b7f-9dcc4d70f647" (UID: "8a311ded-ab56-4f07-8b7f-9dcc4d70f647"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.525457 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5cb8c15-68fc-47db-9f68-1a8401f99f88-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a5cb8c15-68fc-47db-9f68-1a8401f99f88" (UID: "a5cb8c15-68fc-47db-9f68-1a8401f99f88"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.528747 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-kube-api-access-7ttpc" (OuterVolumeSpecName: "kube-api-access-7ttpc") pod "8a311ded-ab56-4f07-8b7f-9dcc4d70f647" (UID: "8a311ded-ab56-4f07-8b7f-9dcc4d70f647"). InnerVolumeSpecName "kube-api-access-7ttpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.533114 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5cb8c15-68fc-47db-9f68-1a8401f99f88-kube-api-access-2hgw6" (OuterVolumeSpecName: "kube-api-access-2hgw6") pod "a5cb8c15-68fc-47db-9f68-1a8401f99f88" (UID: "a5cb8c15-68fc-47db-9f68-1a8401f99f88"). InnerVolumeSpecName "kube-api-access-2hgw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.626210 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hgw6\" (UniqueName: \"kubernetes.io/projected/a5cb8c15-68fc-47db-9f68-1a8401f99f88-kube-api-access-2hgw6\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.626240 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5cb8c15-68fc-47db-9f68-1a8401f99f88-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.626251 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.626262 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ttpc\" (UniqueName: \"kubernetes.io/projected/8a311ded-ab56-4f07-8b7f-9dcc4d70f647-kube-api-access-7ttpc\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.864637 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2rgc6" event={"ID":"a5cb8c15-68fc-47db-9f68-1a8401f99f88","Type":"ContainerDied","Data":"7de096242d5ba0d916d6903858d689e5e9254091b9036d8255334ff0ea8f1f16"} Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.864661 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2rgc6" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.864679 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7de096242d5ba0d916d6903858d689e5e9254091b9036d8255334ff0ea8f1f16" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.866079 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-aecf-account-create-update-gnnvg" event={"ID":"8a311ded-ab56-4f07-8b7f-9dcc4d70f647","Type":"ContainerDied","Data":"04a106975b06c46c735fa10604e40565f439a58073da7467ca0316db88eec4f9"} Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.866113 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04a106975b06c46c735fa10604e40565f439a58073da7467ca0316db88eec4f9" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.866095 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-aecf-account-create-update-gnnvg" Feb 20 08:23:37 crc kubenswrapper[4948]: I0220 08:23:37.868158 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerStarted","Data":"74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05"} Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.025415 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.025809 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.386601 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.535590 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.536147 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.540325 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpq6v\" (UniqueName: \"kubernetes.io/projected/8b167b56-655a-4220-9823-8cc606f5f034-kube-api-access-fpq6v\") pod \"8b167b56-655a-4220-9823-8cc606f5f034\" (UID: \"8b167b56-655a-4220-9823-8cc606f5f034\") " Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.540420 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b167b56-655a-4220-9823-8cc606f5f034-operator-scripts\") pod \"8b167b56-655a-4220-9823-8cc606f5f034\" (UID: \"8b167b56-655a-4220-9823-8cc606f5f034\") " Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.541169 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b167b56-655a-4220-9823-8cc606f5f034-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8b167b56-655a-4220-9823-8cc606f5f034" (UID: "8b167b56-655a-4220-9823-8cc606f5f034"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.549496 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b167b56-655a-4220-9823-8cc606f5f034-kube-api-access-fpq6v" (OuterVolumeSpecName: "kube-api-access-fpq6v") pod "8b167b56-655a-4220-9823-8cc606f5f034" (UID: "8b167b56-655a-4220-9823-8cc606f5f034"). InnerVolumeSpecName "kube-api-access-fpq6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.550713 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.642109 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ssjs\" (UniqueName: \"kubernetes.io/projected/20b61242-b8c1-4e1b-b319-3eafb79ef048-kube-api-access-2ssjs\") pod \"20b61242-b8c1-4e1b-b319-3eafb79ef048\" (UID: \"20b61242-b8c1-4e1b-b319-3eafb79ef048\") " Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.642161 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rx99\" (UniqueName: \"kubernetes.io/projected/fa398740-9a54-472f-9736-0959dd54b657-kube-api-access-7rx99\") pod \"fa398740-9a54-472f-9736-0959dd54b657\" (UID: \"fa398740-9a54-472f-9736-0959dd54b657\") " Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.642249 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl2t2\" (UniqueName: \"kubernetes.io/projected/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-kube-api-access-vl2t2\") pod \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\" (UID: \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\") " Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.642328 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b61242-b8c1-4e1b-b319-3eafb79ef048-operator-scripts\") pod \"20b61242-b8c1-4e1b-b319-3eafb79ef048\" (UID: \"20b61242-b8c1-4e1b-b319-3eafb79ef048\") " Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.642372 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-operator-scripts\") pod \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\" (UID: \"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f\") " Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.642402 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa398740-9a54-472f-9736-0959dd54b657-operator-scripts\") pod \"fa398740-9a54-472f-9736-0959dd54b657\" (UID: \"fa398740-9a54-472f-9736-0959dd54b657\") " Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.643018 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f" (UID: "b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.643445 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa398740-9a54-472f-9736-0959dd54b657-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fa398740-9a54-472f-9736-0959dd54b657" (UID: "fa398740-9a54-472f-9736-0959dd54b657"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.643867 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20b61242-b8c1-4e1b-b319-3eafb79ef048-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "20b61242-b8c1-4e1b-b319-3eafb79ef048" (UID: "20b61242-b8c1-4e1b-b319-3eafb79ef048"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.644201 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpq6v\" (UniqueName: \"kubernetes.io/projected/8b167b56-655a-4220-9823-8cc606f5f034-kube-api-access-fpq6v\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.644216 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b61242-b8c1-4e1b-b319-3eafb79ef048-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.644227 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.644235 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa398740-9a54-472f-9736-0959dd54b657-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.644243 4948 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b167b56-655a-4220-9823-8cc606f5f034-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.644943 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b61242-b8c1-4e1b-b319-3eafb79ef048-kube-api-access-2ssjs" (OuterVolumeSpecName: "kube-api-access-2ssjs") pod "20b61242-b8c1-4e1b-b319-3eafb79ef048" (UID: "20b61242-b8c1-4e1b-b319-3eafb79ef048"). InnerVolumeSpecName "kube-api-access-2ssjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.647424 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-kube-api-access-vl2t2" (OuterVolumeSpecName: "kube-api-access-vl2t2") pod "b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f" (UID: "b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f"). InnerVolumeSpecName "kube-api-access-vl2t2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.648449 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa398740-9a54-472f-9736-0959dd54b657-kube-api-access-7rx99" (OuterVolumeSpecName: "kube-api-access-7rx99") pod "fa398740-9a54-472f-9736-0959dd54b657" (UID: "fa398740-9a54-472f-9736-0959dd54b657"). InnerVolumeSpecName "kube-api-access-7rx99". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.746583 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ssjs\" (UniqueName: \"kubernetes.io/projected/20b61242-b8c1-4e1b-b319-3eafb79ef048-kube-api-access-2ssjs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.747034 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rx99\" (UniqueName: \"kubernetes.io/projected/fa398740-9a54-472f-9736-0959dd54b657-kube-api-access-7rx99\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.747183 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl2t2\" (UniqueName: \"kubernetes.io/projected/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f-kube-api-access-vl2t2\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.878876 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x2pm7" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.878908 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x2pm7" event={"ID":"fa398740-9a54-472f-9736-0959dd54b657","Type":"ContainerDied","Data":"372eccf775e290867aea35db34cd53f0b147aa6e61ff520ef1ac676e72e0326c"} Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.879302 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="372eccf775e290867aea35db34cd53f0b147aa6e61ff520ef1ac676e72e0326c" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.881696 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerStarted","Data":"e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef"} Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.883114 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-mgnlb" event={"ID":"20b61242-b8c1-4e1b-b319-3eafb79ef048","Type":"ContainerDied","Data":"c5c2d7e3e5610297feff1dafc766462b7a8ac2739d1c5be80ff63c9dbf1bcc12"} Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.883132 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-mgnlb" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.883143 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5c2d7e3e5610297feff1dafc766462b7a8ac2739d1c5be80ff63c9dbf1bcc12" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.885059 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" event={"ID":"8b167b56-655a-4220-9823-8cc606f5f034","Type":"ContainerDied","Data":"c9774d1c3be31829d1732d7995a7c4a40d5cd961beaca133cbe7c41c323af95a"} Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.885102 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9774d1c3be31829d1732d7995a7c4a40d5cd961beaca133cbe7c41c323af95a" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.885073 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-78de-account-create-update-fwjdc" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.887018 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" event={"ID":"b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f","Type":"ContainerDied","Data":"4a6359d724bfd5ae6bb40a902ffdabe96c01743e8c81b47b0e26ad235b7eaeff"} Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.887052 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a6359d724bfd5ae6bb40a902ffdabe96c01743e8c81b47b0e26ad235b7eaeff" Feb 20 08:23:38 crc kubenswrapper[4948]: I0220 08:23:38.887092 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ca33-account-create-update-dwqpd" Feb 20 08:23:39 crc kubenswrapper[4948]: I0220 08:23:39.041429 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.905358 4948 generic.go:334] "Generic (PLEG): container finished" podID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerID="577865ca69f0d8457d617948e5cac2be0a124e4620edc60df1395130de8a2334" exitCode=137 Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.905535 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66b4d4cb8-hgmkd" event={"ID":"d1046503-c6d3-49bf-9523-2f897fcb82ba","Type":"ContainerDied","Data":"577865ca69f0d8457d617948e5cac2be0a124e4620edc60df1395130de8a2334"} Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.906091 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66b4d4cb8-hgmkd" event={"ID":"d1046503-c6d3-49bf-9523-2f897fcb82ba","Type":"ContainerDied","Data":"0fc9151c9d2a99b81f2c9346beebf0499347807bacbd7c95da55982c7422c38e"} Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.906110 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fc9151c9d2a99b81f2c9346beebf0499347807bacbd7c95da55982c7422c38e" Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.909132 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerStarted","Data":"52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd"} Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.909284 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="ceilometer-central-agent" containerID="cri-o://ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b" gracePeriod=30 Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.909373 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.909693 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="proxy-httpd" containerID="cri-o://52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd" gracePeriod=30 Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.909743 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="sg-core" containerID="cri-o://e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef" gracePeriod=30 Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.909780 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="ceilometer-notification-agent" containerID="cri-o://74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05" gracePeriod=30 Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.937271 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.963706826 podStartE2EDuration="6.937251168s" podCreationTimestamp="2026-02-20 08:23:34 +0000 UTC" firstStartedPulling="2026-02-20 08:23:35.92817329 +0000 UTC m=+1064.902668110" lastFinishedPulling="2026-02-20 08:23:39.901717632 +0000 UTC m=+1068.876212452" observedRunningTime="2026-02-20 08:23:40.937225837 +0000 UTC m=+1069.911720657" watchObservedRunningTime="2026-02-20 08:23:40.937251168 +0000 UTC m=+1069.911745988" Feb 20 08:23:40 crc kubenswrapper[4948]: I0220 08:23:40.978793 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.097170 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-secret-key\") pod \"d1046503-c6d3-49bf-9523-2f897fcb82ba\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.097237 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgwfv\" (UniqueName: \"kubernetes.io/projected/d1046503-c6d3-49bf-9523-2f897fcb82ba-kube-api-access-mgwfv\") pod \"d1046503-c6d3-49bf-9523-2f897fcb82ba\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.097286 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-tls-certs\") pod \"d1046503-c6d3-49bf-9523-2f897fcb82ba\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.097432 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1046503-c6d3-49bf-9523-2f897fcb82ba-logs\") pod \"d1046503-c6d3-49bf-9523-2f897fcb82ba\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.097456 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-combined-ca-bundle\") pod \"d1046503-c6d3-49bf-9523-2f897fcb82ba\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.097478 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-config-data\") pod \"d1046503-c6d3-49bf-9523-2f897fcb82ba\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.097541 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-scripts\") pod \"d1046503-c6d3-49bf-9523-2f897fcb82ba\" (UID: \"d1046503-c6d3-49bf-9523-2f897fcb82ba\") " Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.098043 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1046503-c6d3-49bf-9523-2f897fcb82ba-logs" (OuterVolumeSpecName: "logs") pod "d1046503-c6d3-49bf-9523-2f897fcb82ba" (UID: "d1046503-c6d3-49bf-9523-2f897fcb82ba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.103148 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d1046503-c6d3-49bf-9523-2f897fcb82ba" (UID: "d1046503-c6d3-49bf-9523-2f897fcb82ba"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.103957 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1046503-c6d3-49bf-9523-2f897fcb82ba-kube-api-access-mgwfv" (OuterVolumeSpecName: "kube-api-access-mgwfv") pod "d1046503-c6d3-49bf-9523-2f897fcb82ba" (UID: "d1046503-c6d3-49bf-9523-2f897fcb82ba"). InnerVolumeSpecName "kube-api-access-mgwfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.124021 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-scripts" (OuterVolumeSpecName: "scripts") pod "d1046503-c6d3-49bf-9523-2f897fcb82ba" (UID: "d1046503-c6d3-49bf-9523-2f897fcb82ba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.127777 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1046503-c6d3-49bf-9523-2f897fcb82ba" (UID: "d1046503-c6d3-49bf-9523-2f897fcb82ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.128788 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-config-data" (OuterVolumeSpecName: "config-data") pod "d1046503-c6d3-49bf-9523-2f897fcb82ba" (UID: "d1046503-c6d3-49bf-9523-2f897fcb82ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.161126 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "d1046503-c6d3-49bf-9523-2f897fcb82ba" (UID: "d1046503-c6d3-49bf-9523-2f897fcb82ba"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.199318 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgwfv\" (UniqueName: \"kubernetes.io/projected/d1046503-c6d3-49bf-9523-2f897fcb82ba-kube-api-access-mgwfv\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.199359 4948 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.199368 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1046503-c6d3-49bf-9523-2f897fcb82ba-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.199377 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.199386 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.199394 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1046503-c6d3-49bf-9523-2f897fcb82ba-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.199402 4948 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d1046503-c6d3-49bf-9523-2f897fcb82ba-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.921388 4948 generic.go:334] "Generic (PLEG): container finished" podID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerID="52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd" exitCode=0 Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.921752 4948 generic.go:334] "Generic (PLEG): container finished" podID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerID="e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef" exitCode=2 Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.921768 4948 generic.go:334] "Generic (PLEG): container finished" podID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerID="74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05" exitCode=0 Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.921554 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerDied","Data":"52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd"} Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.921849 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66b4d4cb8-hgmkd" Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.921895 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerDied","Data":"e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef"} Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.921925 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerDied","Data":"74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05"} Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.961507 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66b4d4cb8-hgmkd"] Feb 20 08:23:41 crc kubenswrapper[4948]: I0220 08:23:41.969336 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-66b4d4cb8-hgmkd"] Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.369549 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.527376 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-config-data\") pod \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.527608 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-scripts\") pod \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.527712 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc4df\" (UniqueName: \"kubernetes.io/projected/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-kube-api-access-qc4df\") pod \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.527767 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-run-httpd\") pod \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.527853 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-log-httpd\") pod \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.527926 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-sg-core-conf-yaml\") pod \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.527996 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-combined-ca-bundle\") pod \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\" (UID: \"eae40be2-f936-4e13-bc8e-0f84bbcd8edd\") " Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.528859 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "eae40be2-f936-4e13-bc8e-0f84bbcd8edd" (UID: "eae40be2-f936-4e13-bc8e-0f84bbcd8edd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.530683 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "eae40be2-f936-4e13-bc8e-0f84bbcd8edd" (UID: "eae40be2-f936-4e13-bc8e-0f84bbcd8edd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.538964 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-kube-api-access-qc4df" (OuterVolumeSpecName: "kube-api-access-qc4df") pod "eae40be2-f936-4e13-bc8e-0f84bbcd8edd" (UID: "eae40be2-f936-4e13-bc8e-0f84bbcd8edd"). InnerVolumeSpecName "kube-api-access-qc4df". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.539450 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-scripts" (OuterVolumeSpecName: "scripts") pod "eae40be2-f936-4e13-bc8e-0f84bbcd8edd" (UID: "eae40be2-f936-4e13-bc8e-0f84bbcd8edd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.559358 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "eae40be2-f936-4e13-bc8e-0f84bbcd8edd" (UID: "eae40be2-f936-4e13-bc8e-0f84bbcd8edd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.624332 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eae40be2-f936-4e13-bc8e-0f84bbcd8edd" (UID: "eae40be2-f936-4e13-bc8e-0f84bbcd8edd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.630034 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.630064 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc4df\" (UniqueName: \"kubernetes.io/projected/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-kube-api-access-qc4df\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.630074 4948 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.630083 4948 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.630092 4948 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.630101 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.666200 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-config-data" (OuterVolumeSpecName: "config-data") pod "eae40be2-f936-4e13-bc8e-0f84bbcd8edd" (UID: "eae40be2-f936-4e13-bc8e-0f84bbcd8edd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.731339 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eae40be2-f936-4e13-bc8e-0f84bbcd8edd-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.756517 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-76b5684765-h5624" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.813994 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b5c5cc45d-7dcbq"] Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.814270 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-b5c5cc45d-7dcbq" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerName="neutron-api" containerID="cri-o://fceb35ea8befbc40c2edb25ea5ef78a9c4a759e83b13e451735a1e2ccf6a99a8" gracePeriod=30 Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.814739 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-b5c5cc45d-7dcbq" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerName="neutron-httpd" containerID="cri-o://4fc7cff150a17fca1672ea7d9eae41c85a175b0d1f41224639307a4d7025cd12" gracePeriod=30 Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.935510 4948 generic.go:334] "Generic (PLEG): container finished" podID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerID="ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b" exitCode=0 Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.935546 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerDied","Data":"ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b"} Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.935570 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eae40be2-f936-4e13-bc8e-0f84bbcd8edd","Type":"ContainerDied","Data":"8b2a81f0ae61bef074923309a4e18638192928b3bbadabe1634088ddab7ac96c"} Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.935586 4948 scope.go:117] "RemoveContainer" containerID="52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.935705 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.966242 4948 scope.go:117] "RemoveContainer" containerID="e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef" Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.969097 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:42 crc kubenswrapper[4948]: I0220 08:23:42.976769 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004307 4948 scope.go:117] "RemoveContainer" containerID="74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004422 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004785 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004800 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004811 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon-log" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004817 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon-log" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004826 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b167b56-655a-4220-9823-8cc606f5f034" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004833 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b167b56-655a-4220-9823-8cc606f5f034" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004841 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="ceilometer-central-agent" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004847 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="ceilometer-central-agent" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004856 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="ceilometer-notification-agent" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004863 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="ceilometer-notification-agent" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004877 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a311ded-ab56-4f07-8b7f-9dcc4d70f647" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004883 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a311ded-ab56-4f07-8b7f-9dcc4d70f647" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004899 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="proxy-httpd" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004905 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="proxy-httpd" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004921 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5cb8c15-68fc-47db-9f68-1a8401f99f88" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004927 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5cb8c15-68fc-47db-9f68-1a8401f99f88" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004935 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="sg-core" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004942 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="sg-core" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004950 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20b61242-b8c1-4e1b-b319-3eafb79ef048" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.004956 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="20b61242-b8c1-4e1b-b319-3eafb79ef048" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.004964 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.005016 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.005028 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa398740-9a54-472f-9736-0959dd54b657" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.005035 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa398740-9a54-472f-9736-0959dd54b657" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006376 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="ceilometer-notification-agent" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006400 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon-log" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006413 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="20b61242-b8c1-4e1b-b319-3eafb79ef048" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006423 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b167b56-655a-4220-9823-8cc606f5f034" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006445 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a311ded-ab56-4f07-8b7f-9dcc4d70f647" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006457 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="proxy-httpd" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006470 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f" containerName="mariadb-account-create-update" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006483 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5cb8c15-68fc-47db-9f68-1a8401f99f88" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006498 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="ceilometer-central-agent" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006507 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa398740-9a54-472f-9736-0959dd54b657" containerName="mariadb-database-create" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006521 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" containerName="sg-core" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.006532 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" containerName="horizon" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.008566 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.012347 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.012572 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.033769 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.040108 4948 scope.go:117] "RemoveContainer" containerID="ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.096330 4948 scope.go:117] "RemoveContainer" containerID="52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.096684 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd\": container with ID starting with 52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd not found: ID does not exist" containerID="52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.096718 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd"} err="failed to get container status \"52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd\": rpc error: code = NotFound desc = could not find container \"52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd\": container with ID starting with 52810114bb238a5ef2002a9ed77afb4259619263e919c688def5a10a52ac2cbd not found: ID does not exist" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.096738 4948 scope.go:117] "RemoveContainer" containerID="e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.096940 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef\": container with ID starting with e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef not found: ID does not exist" containerID="e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.096963 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef"} err="failed to get container status \"e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef\": rpc error: code = NotFound desc = could not find container \"e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef\": container with ID starting with e45d96dd2c40e8be3085b92f36d19f6573863121a02ff440da9b19f454a42aef not found: ID does not exist" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.096990 4948 scope.go:117] "RemoveContainer" containerID="74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.097283 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05\": container with ID starting with 74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05 not found: ID does not exist" containerID="74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.097321 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05"} err="failed to get container status \"74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05\": rpc error: code = NotFound desc = could not find container \"74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05\": container with ID starting with 74627264af113298ec39e3890cccb5271e59d7d4c945fe853668a4d33ea1db05 not found: ID does not exist" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.097345 4948 scope.go:117] "RemoveContainer" containerID="ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b" Feb 20 08:23:43 crc kubenswrapper[4948]: E0220 08:23:43.097551 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b\": container with ID starting with ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b not found: ID does not exist" containerID="ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.097571 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b"} err="failed to get container status \"ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b\": rpc error: code = NotFound desc = could not find container \"ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b\": container with ID starting with ef2c6dbff693c9b5a07ebf11c4675349807817c4344b76c4fc763563e5428a3b not found: ID does not exist" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.149118 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-config-data\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.149182 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txt8k\" (UniqueName: \"kubernetes.io/projected/0894350a-0df8-4c04-b7c7-fd8f93139959-kube-api-access-txt8k\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.149206 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-log-httpd\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.149386 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.149444 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-run-httpd\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.149604 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-scripts\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.149687 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.251606 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-scripts\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.251666 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.251705 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-config-data\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.251747 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txt8k\" (UniqueName: \"kubernetes.io/projected/0894350a-0df8-4c04-b7c7-fd8f93139959-kube-api-access-txt8k\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.251775 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-log-httpd\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.251814 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.251834 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-run-httpd\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.252308 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-run-httpd\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.252441 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-log-httpd\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.256737 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-scripts\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.257777 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-config-data\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.258517 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.259321 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.277342 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txt8k\" (UniqueName: \"kubernetes.io/projected/0894350a-0df8-4c04-b7c7-fd8f93139959-kube-api-access-txt8k\") pod \"ceilometer-0\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.332020 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.736277 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1046503-c6d3-49bf-9523-2f897fcb82ba" path="/var/lib/kubelet/pods/d1046503-c6d3-49bf-9523-2f897fcb82ba/volumes" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.737226 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eae40be2-f936-4e13-bc8e-0f84bbcd8edd" path="/var/lib/kubelet/pods/eae40be2-f936-4e13-bc8e-0f84bbcd8edd/volumes" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.813709 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:43 crc kubenswrapper[4948]: W0220 08:23:43.818395 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0894350a_0df8_4c04_b7c7_fd8f93139959.slice/crio-6657eaf082a861ad8f0f29c4c00aaf35d258d94004b82734a54c5722b14133f5 WatchSource:0}: Error finding container 6657eaf082a861ad8f0f29c4c00aaf35d258d94004b82734a54c5722b14133f5: Status 404 returned error can't find the container with id 6657eaf082a861ad8f0f29c4c00aaf35d258d94004b82734a54c5722b14133f5 Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.946620 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerStarted","Data":"6657eaf082a861ad8f0f29c4c00aaf35d258d94004b82734a54c5722b14133f5"} Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.952506 4948 generic.go:334] "Generic (PLEG): container finished" podID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerID="4fc7cff150a17fca1672ea7d9eae41c85a175b0d1f41224639307a4d7025cd12" exitCode=0 Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.952590 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5c5cc45d-7dcbq" event={"ID":"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d","Type":"ContainerDied","Data":"4fc7cff150a17fca1672ea7d9eae41c85a175b0d1f41224639307a4d7025cd12"} Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.968134 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:43 crc kubenswrapper[4948]: I0220 08:23:43.973728 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7c7cd9876c-262qf" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.518730 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xt5t7"] Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.519859 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.523657 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.523785 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2tb2s" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.523933 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.573606 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xt5t7"] Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.677393 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-config-data\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.677516 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4jh4\" (UniqueName: \"kubernetes.io/projected/3fdb2e02-8713-4205-ae31-0af33f8e4047-kube-api-access-j4jh4\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.677542 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.677559 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-scripts\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.780116 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-config-data\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.781401 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4jh4\" (UniqueName: \"kubernetes.io/projected/3fdb2e02-8713-4205-ae31-0af33f8e4047-kube-api-access-j4jh4\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.781450 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.781483 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-scripts\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.786813 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-scripts\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.786938 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-config-data\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.788651 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.800822 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4jh4\" (UniqueName: \"kubernetes.io/projected/3fdb2e02-8713-4205-ae31-0af33f8e4047-kube-api-access-j4jh4\") pod \"nova-cell0-conductor-db-sync-xt5t7\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:44 crc kubenswrapper[4948]: I0220 08:23:44.844755 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:23:45 crc kubenswrapper[4948]: I0220 08:23:45.302892 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xt5t7"] Feb 20 08:23:45 crc kubenswrapper[4948]: I0220 08:23:45.984648 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" event={"ID":"3fdb2e02-8713-4205-ae31-0af33f8e4047","Type":"ContainerStarted","Data":"d20f6a9cd2d1756d07c023cf4fc3e418531e5c20161aa674031ec3aa6a8df0d2"} Feb 20 08:23:46 crc kubenswrapper[4948]: I0220 08:23:46.911850 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:23:46 crc kubenswrapper[4948]: I0220 08:23:46.995248 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerStarted","Data":"8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc"} Feb 20 08:23:46 crc kubenswrapper[4948]: I0220 08:23:46.995286 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerStarted","Data":"5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4"} Feb 20 08:23:48 crc kubenswrapper[4948]: I0220 08:23:48.016249 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerStarted","Data":"d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed"} Feb 20 08:23:49 crc kubenswrapper[4948]: I0220 08:23:49.027062 4948 generic.go:334] "Generic (PLEG): container finished" podID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerID="fceb35ea8befbc40c2edb25ea5ef78a9c4a759e83b13e451735a1e2ccf6a99a8" exitCode=0 Feb 20 08:23:49 crc kubenswrapper[4948]: I0220 08:23:49.027148 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5c5cc45d-7dcbq" event={"ID":"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d","Type":"ContainerDied","Data":"fceb35ea8befbc40c2edb25ea5ef78a9c4a759e83b13e451735a1e2ccf6a99a8"} Feb 20 08:23:51 crc kubenswrapper[4948]: I0220 08:23:51.350279 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:51 crc kubenswrapper[4948]: I0220 08:23:51.365393 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-86887846d-pqjcb" Feb 20 08:23:51 crc kubenswrapper[4948]: I0220 08:23:51.427321 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-d46988dcb-rn6qs"] Feb 20 08:23:51 crc kubenswrapper[4948]: I0220 08:23:51.427864 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-d46988dcb-rn6qs" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerName="placement-log" containerID="cri-o://26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26" gracePeriod=30 Feb 20 08:23:51 crc kubenswrapper[4948]: I0220 08:23:51.428021 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-d46988dcb-rn6qs" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerName="placement-api" containerID="cri-o://6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af" gracePeriod=30 Feb 20 08:23:52 crc kubenswrapper[4948]: I0220 08:23:52.072119 4948 generic.go:334] "Generic (PLEG): container finished" podID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerID="26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26" exitCode=143 Feb 20 08:23:52 crc kubenswrapper[4948]: I0220 08:23:52.072205 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d46988dcb-rn6qs" event={"ID":"8b1ee525-1a6e-461d-aa4d-178ab601ec9d","Type":"ContainerDied","Data":"26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26"} Feb 20 08:23:52 crc kubenswrapper[4948]: I0220 08:23:52.406201 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:23:52 crc kubenswrapper[4948]: I0220 08:23:52.406428 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerName="glance-log" containerID="cri-o://a66ddb00a7f57dc492a3a8a03a3bce06f45c65381b1e5db4f923e10fec5267a0" gracePeriod=30 Feb 20 08:23:52 crc kubenswrapper[4948]: I0220 08:23:52.406545 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerName="glance-httpd" containerID="cri-o://656e84496ca041e8a454b07ed67e165b130af1030016564ae4519b8e2e6f11e5" gracePeriod=30 Feb 20 08:23:53 crc kubenswrapper[4948]: I0220 08:23:53.084216 4948 generic.go:334] "Generic (PLEG): container finished" podID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerID="a66ddb00a7f57dc492a3a8a03a3bce06f45c65381b1e5db4f923e10fec5267a0" exitCode=143 Feb 20 08:23:53 crc kubenswrapper[4948]: I0220 08:23:53.084249 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d9399b9c-0935-4735-8b08-96d4a29d4ba8","Type":"ContainerDied","Data":"a66ddb00a7f57dc492a3a8a03a3bce06f45c65381b1e5db4f923e10fec5267a0"} Feb 20 08:23:53 crc kubenswrapper[4948]: I0220 08:23:53.225834 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:23:53 crc kubenswrapper[4948]: I0220 08:23:53.226161 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" containerName="glance-log" containerID="cri-o://9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166" gracePeriod=30 Feb 20 08:23:53 crc kubenswrapper[4948]: I0220 08:23:53.226626 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" containerName="glance-httpd" containerID="cri-o://08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679" gracePeriod=30 Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.107849 4948 generic.go:334] "Generic (PLEG): container finished" podID="33b35605-1057-4e64-99a7-a7273351d6c2" containerID="9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166" exitCode=143 Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.107948 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33b35605-1057-4e64-99a7-a7273351d6c2","Type":"ContainerDied","Data":"9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166"} Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.113846 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="proxy-httpd" containerID="cri-o://c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5" gracePeriod=30 Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.113877 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.113843 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="ceilometer-central-agent" containerID="cri-o://5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4" gracePeriod=30 Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.113936 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="sg-core" containerID="cri-o://d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed" gracePeriod=30 Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.113943 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="ceilometer-notification-agent" containerID="cri-o://8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc" gracePeriod=30 Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.114195 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.115648 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" event={"ID":"3fdb2e02-8713-4205-ae31-0af33f8e4047","Type":"ContainerStarted","Data":"53a1cca1717eaf7448d7d84b4b68c8bad0f23b2d1cc07d696f16d15b417b8356"} Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.152638 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.105361799 podStartE2EDuration="12.152620418s" podCreationTimestamp="2026-02-20 08:23:42 +0000 UTC" firstStartedPulling="2026-02-20 08:23:43.821928265 +0000 UTC m=+1072.796423085" lastFinishedPulling="2026-02-20 08:23:53.869186884 +0000 UTC m=+1082.843681704" observedRunningTime="2026-02-20 08:23:54.144107657 +0000 UTC m=+1083.118602477" watchObservedRunningTime="2026-02-20 08:23:54.152620418 +0000 UTC m=+1083.127115238" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.157730 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-ovndb-tls-certs\") pod \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.157846 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-combined-ca-bundle\") pod \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.157935 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khls2\" (UniqueName: \"kubernetes.io/projected/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-kube-api-access-khls2\") pod \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.157989 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-httpd-config\") pod \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.158137 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-config\") pod \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\" (UID: \"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d\") " Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.166238 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" (UID: "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.166487 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-kube-api-access-khls2" (OuterVolumeSpecName: "kube-api-access-khls2") pod "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" (UID: "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d"). InnerVolumeSpecName "kube-api-access-khls2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.199279 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" podStartSLOduration=1.6380877090000001 podStartE2EDuration="10.199256802s" podCreationTimestamp="2026-02-20 08:23:44 +0000 UTC" firstStartedPulling="2026-02-20 08:23:45.30840186 +0000 UTC m=+1074.282896710" lastFinishedPulling="2026-02-20 08:23:53.869570983 +0000 UTC m=+1082.844065803" observedRunningTime="2026-02-20 08:23:54.186610499 +0000 UTC m=+1083.161105319" watchObservedRunningTime="2026-02-20 08:23:54.199256802 +0000 UTC m=+1083.173751622" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.214405 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-config" (OuterVolumeSpecName: "config") pod "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" (UID: "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.230842 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" (UID: "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.254407 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" (UID: "fd1b85e2-1b5f-4993-bc66-72daaf6aa79d"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.260773 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.260802 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khls2\" (UniqueName: \"kubernetes.io/projected/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-kube-api-access-khls2\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.260812 4948 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.260821 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:54 crc kubenswrapper[4948]: I0220 08:23:54.260830 4948 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.098983 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.125617 4948 generic.go:334] "Generic (PLEG): container finished" podID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerID="6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af" exitCode=0 Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.125676 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d46988dcb-rn6qs" event={"ID":"8b1ee525-1a6e-461d-aa4d-178ab601ec9d","Type":"ContainerDied","Data":"6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af"} Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.125704 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d46988dcb-rn6qs" event={"ID":"8b1ee525-1a6e-461d-aa4d-178ab601ec9d","Type":"ContainerDied","Data":"101e3a4bd62000e7fe0c14babf5f059dc5dad0bc38ccadaa7bd0301c2ad9ec7f"} Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.125720 4948 scope.go:117] "RemoveContainer" containerID="6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.125836 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d46988dcb-rn6qs" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.146238 4948 generic.go:334] "Generic (PLEG): container finished" podID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerID="d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed" exitCode=2 Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.146267 4948 generic.go:334] "Generic (PLEG): container finished" podID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerID="8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc" exitCode=0 Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.146334 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerStarted","Data":"c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5"} Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.146359 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerDied","Data":"d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed"} Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.146371 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerDied","Data":"8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc"} Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.149305 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b5c5cc45d-7dcbq" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.149716 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5c5cc45d-7dcbq" event={"ID":"fd1b85e2-1b5f-4993-bc66-72daaf6aa79d","Type":"ContainerDied","Data":"4c4522a0ae94ffaed4b0ddc1a755c0498d0a11f95af5418408b70c0351708567"} Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.170210 4948 scope.go:117] "RemoveContainer" containerID="26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.176582 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-public-tls-certs\") pod \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.176650 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-internal-tls-certs\") pod \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.176695 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-logs\") pod \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.176775 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-scripts\") pod \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.176848 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bpmf\" (UniqueName: \"kubernetes.io/projected/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-kube-api-access-2bpmf\") pod \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.176937 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-combined-ca-bundle\") pod \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.176998 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-config-data\") pod \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\" (UID: \"8b1ee525-1a6e-461d-aa4d-178ab601ec9d\") " Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.177876 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-logs" (OuterVolumeSpecName: "logs") pod "8b1ee525-1a6e-461d-aa4d-178ab601ec9d" (UID: "8b1ee525-1a6e-461d-aa4d-178ab601ec9d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.185490 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-scripts" (OuterVolumeSpecName: "scripts") pod "8b1ee525-1a6e-461d-aa4d-178ab601ec9d" (UID: "8b1ee525-1a6e-461d-aa4d-178ab601ec9d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.187892 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b5c5cc45d-7dcbq"] Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.200081 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-kube-api-access-2bpmf" (OuterVolumeSpecName: "kube-api-access-2bpmf") pod "8b1ee525-1a6e-461d-aa4d-178ab601ec9d" (UID: "8b1ee525-1a6e-461d-aa4d-178ab601ec9d"). InnerVolumeSpecName "kube-api-access-2bpmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.206452 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b5c5cc45d-7dcbq"] Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.269503 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8b1ee525-1a6e-461d-aa4d-178ab601ec9d" (UID: "8b1ee525-1a6e-461d-aa4d-178ab601ec9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.282617 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.282642 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bpmf\" (UniqueName: \"kubernetes.io/projected/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-kube-api-access-2bpmf\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.282653 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.282662 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.333607 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-config-data" (OuterVolumeSpecName: "config-data") pod "8b1ee525-1a6e-461d-aa4d-178ab601ec9d" (UID: "8b1ee525-1a6e-461d-aa4d-178ab601ec9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.338619 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8b1ee525-1a6e-461d-aa4d-178ab601ec9d" (UID: "8b1ee525-1a6e-461d-aa4d-178ab601ec9d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.339113 4948 scope.go:117] "RemoveContainer" containerID="6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af" Feb 20 08:23:55 crc kubenswrapper[4948]: E0220 08:23:55.339870 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af\": container with ID starting with 6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af not found: ID does not exist" containerID="6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.339956 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af"} err="failed to get container status \"6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af\": rpc error: code = NotFound desc = could not find container \"6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af\": container with ID starting with 6a2830020d67b39ba1edbe811de2ebbd05b2b8a02d0578d95553d45d985544af not found: ID does not exist" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.340078 4948 scope.go:117] "RemoveContainer" containerID="26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26" Feb 20 08:23:55 crc kubenswrapper[4948]: E0220 08:23:55.344104 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26\": container with ID starting with 26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26 not found: ID does not exist" containerID="26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.344223 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26"} err="failed to get container status \"26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26\": rpc error: code = NotFound desc = could not find container \"26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26\": container with ID starting with 26cbea34a0a499d9db14acd1bf75d24dc0d0a677df528028730bacad671ebd26 not found: ID does not exist" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.344294 4948 scope.go:117] "RemoveContainer" containerID="4fc7cff150a17fca1672ea7d9eae41c85a175b0d1f41224639307a4d7025cd12" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.346893 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8b1ee525-1a6e-461d-aa4d-178ab601ec9d" (UID: "8b1ee525-1a6e-461d-aa4d-178ab601ec9d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.371835 4948 scope.go:117] "RemoveContainer" containerID="fceb35ea8befbc40c2edb25ea5ef78a9c4a759e83b13e451735a1e2ccf6a99a8" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.384276 4948 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.384423 4948 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.384478 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b1ee525-1a6e-461d-aa4d-178ab601ec9d-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.456906 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-d46988dcb-rn6qs"] Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.469518 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-d46988dcb-rn6qs"] Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.733074 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" path="/var/lib/kubelet/pods/8b1ee525-1a6e-461d-aa4d-178ab601ec9d/volumes" Feb 20 08:23:55 crc kubenswrapper[4948]: I0220 08:23:55.733823 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" path="/var/lib/kubelet/pods/fd1b85e2-1b5f-4993-bc66-72daaf6aa79d/volumes" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.157291 4948 generic.go:334] "Generic (PLEG): container finished" podID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerID="656e84496ca041e8a454b07ed67e165b130af1030016564ae4519b8e2e6f11e5" exitCode=0 Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.157369 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d9399b9c-0935-4735-8b08-96d4a29d4ba8","Type":"ContainerDied","Data":"656e84496ca041e8a454b07ed67e165b130af1030016564ae4519b8e2e6f11e5"} Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.157555 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d9399b9c-0935-4735-8b08-96d4a29d4ba8","Type":"ContainerDied","Data":"fecc6864756af4b1a18a9eaf7ed69fde086ccbc2685b82cc1f598b9178cb31a0"} Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.157570 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fecc6864756af4b1a18a9eaf7ed69fde086ccbc2685b82cc1f598b9178cb31a0" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.165045 4948 generic.go:334] "Generic (PLEG): container finished" podID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerID="5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4" exitCode=0 Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.165079 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerDied","Data":"5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4"} Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.181308 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.302243 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-public-tls-certs\") pod \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.302302 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdk7s\" (UniqueName: \"kubernetes.io/projected/d9399b9c-0935-4735-8b08-96d4a29d4ba8-kube-api-access-jdk7s\") pod \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.302345 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-logs\") pod \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.302405 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-combined-ca-bundle\") pod \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.302481 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-httpd-run\") pod \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.302564 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.302625 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-config-data\") pod \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.302654 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-scripts\") pod \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\" (UID: \"d9399b9c-0935-4735-8b08-96d4a29d4ba8\") " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.303364 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-logs" (OuterVolumeSpecName: "logs") pod "d9399b9c-0935-4735-8b08-96d4a29d4ba8" (UID: "d9399b9c-0935-4735-8b08-96d4a29d4ba8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.303386 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d9399b9c-0935-4735-8b08-96d4a29d4ba8" (UID: "d9399b9c-0935-4735-8b08-96d4a29d4ba8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.306765 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9399b9c-0935-4735-8b08-96d4a29d4ba8-kube-api-access-jdk7s" (OuterVolumeSpecName: "kube-api-access-jdk7s") pod "d9399b9c-0935-4735-8b08-96d4a29d4ba8" (UID: "d9399b9c-0935-4735-8b08-96d4a29d4ba8"). InnerVolumeSpecName "kube-api-access-jdk7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.324115 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "d9399b9c-0935-4735-8b08-96d4a29d4ba8" (UID: "d9399b9c-0935-4735-8b08-96d4a29d4ba8"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.324200 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-scripts" (OuterVolumeSpecName: "scripts") pod "d9399b9c-0935-4735-8b08-96d4a29d4ba8" (UID: "d9399b9c-0935-4735-8b08-96d4a29d4ba8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.332938 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d9399b9c-0935-4735-8b08-96d4a29d4ba8" (UID: "d9399b9c-0935-4735-8b08-96d4a29d4ba8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.357626 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d9399b9c-0935-4735-8b08-96d4a29d4ba8" (UID: "d9399b9c-0935-4735-8b08-96d4a29d4ba8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.393595 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-config-data" (OuterVolumeSpecName: "config-data") pod "d9399b9c-0935-4735-8b08-96d4a29d4ba8" (UID: "d9399b9c-0935-4735-8b08-96d4a29d4ba8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.405035 4948 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.405105 4948 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.405118 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.405129 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.405140 4948 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.405232 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdk7s\" (UniqueName: \"kubernetes.io/projected/d9399b9c-0935-4735-8b08-96d4a29d4ba8-kube-api-access-jdk7s\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.405246 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9399b9c-0935-4735-8b08-96d4a29d4ba8-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.405257 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9399b9c-0935-4735-8b08-96d4a29d4ba8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.425173 4948 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.506776 4948 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:56 crc kubenswrapper[4948]: I0220 08:23:56.903604 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.016940 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-internal-tls-certs\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.017017 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-httpd-run\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.017045 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-combined-ca-bundle\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.017099 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-config-data\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.017114 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-logs\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.017134 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-scripts\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.017193 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6phw\" (UniqueName: \"kubernetes.io/projected/33b35605-1057-4e64-99a7-a7273351d6c2-kube-api-access-q6phw\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.017231 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.019185 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.027089 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-logs" (OuterVolumeSpecName: "logs") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.037343 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.039452 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-scripts" (OuterVolumeSpecName: "scripts") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.066170 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33b35605-1057-4e64-99a7-a7273351d6c2-kube-api-access-q6phw" (OuterVolumeSpecName: "kube-api-access-q6phw") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "kube-api-access-q6phw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.103795 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.118218 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-config-data" (OuterVolumeSpecName: "config-data") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.118726 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-config-data\") pod \"33b35605-1057-4e64-99a7-a7273351d6c2\" (UID: \"33b35605-1057-4e64-99a7-a7273351d6c2\") " Feb 20 08:23:57 crc kubenswrapper[4948]: W0220 08:23:57.119183 4948 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/33b35605-1057-4e64-99a7-a7273351d6c2/volumes/kubernetes.io~secret/config-data Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.122708 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-config-data" (OuterVolumeSpecName: "config-data") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.119913 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "33b35605-1057-4e64-99a7-a7273351d6c2" (UID: "33b35605-1057-4e64-99a7-a7273351d6c2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.119333 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6phw\" (UniqueName: \"kubernetes.io/projected/33b35605-1057-4e64-99a7-a7273351d6c2-kube-api-access-q6phw\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.123144 4948 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.127625 4948 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.127893 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.128021 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.128102 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33b35605-1057-4e64-99a7-a7273351d6c2-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.128158 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.145927 4948 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.180370 4948 generic.go:334] "Generic (PLEG): container finished" podID="33b35605-1057-4e64-99a7-a7273351d6c2" containerID="08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679" exitCode=0 Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.180502 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.182029 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.183020 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33b35605-1057-4e64-99a7-a7273351d6c2","Type":"ContainerDied","Data":"08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679"} Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.183123 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33b35605-1057-4e64-99a7-a7273351d6c2","Type":"ContainerDied","Data":"3abd35d213010021a69fde90dd757e9181aec2b84f4e7f71376ea0edef4300a5"} Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.183150 4948 scope.go:117] "RemoveContainer" containerID="08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.221163 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.229882 4948 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33b35605-1057-4e64-99a7-a7273351d6c2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.229940 4948 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.236334 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.246915 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.261105 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284290 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.284646 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" containerName="glance-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284658 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" containerName="glance-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.284669 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerName="neutron-api" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284676 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerName="neutron-api" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.284690 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerName="placement-api" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284697 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerName="placement-api" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.284718 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerName="neutron-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284724 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerName="neutron-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.284742 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerName="placement-log" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284747 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerName="placement-log" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.284757 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerName="glance-log" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284762 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerName="glance-log" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.284774 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" containerName="glance-log" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284779 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" containerName="glance-log" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.284792 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerName="glance-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284799 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerName="glance-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.284989 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" containerName="glance-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.285011 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerName="placement-log" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.285020 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerName="glance-log" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.285031 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" containerName="glance-log" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.285041 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerName="neutron-api" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.285047 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd1b85e2-1b5f-4993-bc66-72daaf6aa79d" containerName="neutron-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.285056 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b1ee525-1a6e-461d-aa4d-178ab601ec9d" containerName="placement-api" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.285067 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" containerName="glance-httpd" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.285858 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.286804 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.286879 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.287373 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.300658 4948 scope.go:117] "RemoveContainer" containerID="9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.308117 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.308351 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.308588 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.308716 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.308932 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.309242 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bgxj8" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331442 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-config-data\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331474 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331494 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331523 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4fcb965e-4d04-4863-a966-39a83f458fa6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331541 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-scripts\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331558 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331578 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-logs\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331607 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331624 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331644 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331688 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx7wl\" (UniqueName: \"kubernetes.io/projected/4fcb965e-4d04-4863-a966-39a83f458fa6-kube-api-access-nx7wl\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331708 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl9fz\" (UniqueName: \"kubernetes.io/projected/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-kube-api-access-xl9fz\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331731 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331758 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331798 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.331818 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fcb965e-4d04-4863-a966-39a83f458fa6-logs\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.357731 4948 scope.go:117] "RemoveContainer" containerID="08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.358430 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679\": container with ID starting with 08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679 not found: ID does not exist" containerID="08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.358492 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679"} err="failed to get container status \"08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679\": rpc error: code = NotFound desc = could not find container \"08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679\": container with ID starting with 08ec8e2a4374c5a50916bcc6e4fbc0d7957aac63564273dd75ed243c1b2a0679 not found: ID does not exist" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.358532 4948 scope.go:117] "RemoveContainer" containerID="9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166" Feb 20 08:23:57 crc kubenswrapper[4948]: E0220 08:23:57.358861 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166\": container with ID starting with 9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166 not found: ID does not exist" containerID="9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.358900 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166"} err="failed to get container status \"9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166\": rpc error: code = NotFound desc = could not find container \"9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166\": container with ID starting with 9232d9b3b9f362e1b01dbc32c5300081164bdbed5e40b2c4cd59a5e4f4b28166 not found: ID does not exist" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.382738 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.433778 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fcb965e-4d04-4863-a966-39a83f458fa6-logs\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.433843 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-config-data\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.433870 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.433893 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.433934 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4fcb965e-4d04-4863-a966-39a83f458fa6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.433956 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-scripts\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434021 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434046 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-logs\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434073 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434094 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434123 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434193 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx7wl\" (UniqueName: \"kubernetes.io/projected/4fcb965e-4d04-4863-a966-39a83f458fa6-kube-api-access-nx7wl\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434218 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl9fz\" (UniqueName: \"kubernetes.io/projected/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-kube-api-access-xl9fz\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434261 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434304 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434323 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434340 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434259 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fcb965e-4d04-4863-a966-39a83f458fa6-logs\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434675 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4fcb965e-4d04-4863-a966-39a83f458fa6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.434684 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-logs\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.435041 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.435442 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.438765 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.439007 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.439998 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.440851 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-config-data\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.442108 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.442388 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.453240 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fcb965e-4d04-4863-a966-39a83f458fa6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.453706 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-scripts\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.457529 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl9fz\" (UniqueName: \"kubernetes.io/projected/acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a-kube-api-access-xl9fz\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.457922 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx7wl\" (UniqueName: \"kubernetes.io/projected/4fcb965e-4d04-4863-a966-39a83f458fa6-kube-api-access-nx7wl\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.475173 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a\") " pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.486826 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"4fcb965e-4d04-4863-a966-39a83f458fa6\") " pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.623237 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.637440 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.738709 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33b35605-1057-4e64-99a7-a7273351d6c2" path="/var/lib/kubelet/pods/33b35605-1057-4e64-99a7-a7273351d6c2/volumes" Feb 20 08:23:57 crc kubenswrapper[4948]: I0220 08:23:57.739904 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9399b9c-0935-4735-8b08-96d4a29d4ba8" path="/var/lib/kubelet/pods/d9399b9c-0935-4735-8b08-96d4a29d4ba8/volumes" Feb 20 08:23:58 crc kubenswrapper[4948]: I0220 08:23:58.206620 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 08:23:58 crc kubenswrapper[4948]: I0220 08:23:58.332536 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 08:23:58 crc kubenswrapper[4948]: W0220 08:23:58.342557 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podacd5bc1f_fee5_4ef1_95f9_5c4d11bbc54a.slice/crio-584190f9085e41c64b0b41271b19963599bd1689d31b9325f6e15ec1fa547bb8 WatchSource:0}: Error finding container 584190f9085e41c64b0b41271b19963599bd1689d31b9325f6e15ec1fa547bb8: Status 404 returned error can't find the container with id 584190f9085e41c64b0b41271b19963599bd1689d31b9325f6e15ec1fa547bb8 Feb 20 08:23:59 crc kubenswrapper[4948]: I0220 08:23:59.203165 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4fcb965e-4d04-4863-a966-39a83f458fa6","Type":"ContainerStarted","Data":"d76f6785864615fbf907238c357a5d56e0a2c4b94dbff43d8df27025c8acb1e5"} Feb 20 08:23:59 crc kubenswrapper[4948]: I0220 08:23:59.208787 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a","Type":"ContainerStarted","Data":"584190f9085e41c64b0b41271b19963599bd1689d31b9325f6e15ec1fa547bb8"} Feb 20 08:24:00 crc kubenswrapper[4948]: I0220 08:24:00.217290 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4fcb965e-4d04-4863-a966-39a83f458fa6","Type":"ContainerStarted","Data":"d4b2de7726d17bd80f2876aa816146886b795551e9fe8200d4594e8816986240"} Feb 20 08:24:00 crc kubenswrapper[4948]: I0220 08:24:00.217735 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4fcb965e-4d04-4863-a966-39a83f458fa6","Type":"ContainerStarted","Data":"faa90aa4e853b5ab666d9f1be43146101722dceedf35406ce1ae282e0648e2d0"} Feb 20 08:24:00 crc kubenswrapper[4948]: I0220 08:24:00.220134 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a","Type":"ContainerStarted","Data":"d5c3a0258f4fedb1c452016b5c3dfafb6d6a2a6693dd942cded6707663c562d1"} Feb 20 08:24:00 crc kubenswrapper[4948]: I0220 08:24:00.220157 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a","Type":"ContainerStarted","Data":"168c4e34c8fa1f4e36ab74b6fd468244e8fe2da745463fcd00309d5e31d08129"} Feb 20 08:24:00 crc kubenswrapper[4948]: I0220 08:24:00.239677 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.239660963 podStartE2EDuration="3.239660963s" podCreationTimestamp="2026-02-20 08:23:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:00.236395852 +0000 UTC m=+1089.210890662" watchObservedRunningTime="2026-02-20 08:24:00.239660963 +0000 UTC m=+1089.214155783" Feb 20 08:24:00 crc kubenswrapper[4948]: I0220 08:24:00.264589 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.2645723 podStartE2EDuration="3.2645723s" podCreationTimestamp="2026-02-20 08:23:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:00.259476044 +0000 UTC m=+1089.233970864" watchObservedRunningTime="2026-02-20 08:24:00.2645723 +0000 UTC m=+1089.239067120" Feb 20 08:24:06 crc kubenswrapper[4948]: I0220 08:24:06.272140 4948 generic.go:334] "Generic (PLEG): container finished" podID="3fdb2e02-8713-4205-ae31-0af33f8e4047" containerID="53a1cca1717eaf7448d7d84b4b68c8bad0f23b2d1cc07d696f16d15b417b8356" exitCode=0 Feb 20 08:24:06 crc kubenswrapper[4948]: I0220 08:24:06.272310 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" event={"ID":"3fdb2e02-8713-4205-ae31-0af33f8e4047","Type":"ContainerDied","Data":"53a1cca1717eaf7448d7d84b4b68c8bad0f23b2d1cc07d696f16d15b417b8356"} Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.623930 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.624295 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.641265 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.641701 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.655346 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.665027 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.671596 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.692057 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.722896 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.845822 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-scripts\") pod \"3fdb2e02-8713-4205-ae31-0af33f8e4047\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.845873 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-config-data\") pod \"3fdb2e02-8713-4205-ae31-0af33f8e4047\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.846009 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4jh4\" (UniqueName: \"kubernetes.io/projected/3fdb2e02-8713-4205-ae31-0af33f8e4047-kube-api-access-j4jh4\") pod \"3fdb2e02-8713-4205-ae31-0af33f8e4047\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.846095 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-combined-ca-bundle\") pod \"3fdb2e02-8713-4205-ae31-0af33f8e4047\" (UID: \"3fdb2e02-8713-4205-ae31-0af33f8e4047\") " Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.853149 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-scripts" (OuterVolumeSpecName: "scripts") pod "3fdb2e02-8713-4205-ae31-0af33f8e4047" (UID: "3fdb2e02-8713-4205-ae31-0af33f8e4047"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.853171 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fdb2e02-8713-4205-ae31-0af33f8e4047-kube-api-access-j4jh4" (OuterVolumeSpecName: "kube-api-access-j4jh4") pod "3fdb2e02-8713-4205-ae31-0af33f8e4047" (UID: "3fdb2e02-8713-4205-ae31-0af33f8e4047"). InnerVolumeSpecName "kube-api-access-j4jh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.871606 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-config-data" (OuterVolumeSpecName: "config-data") pod "3fdb2e02-8713-4205-ae31-0af33f8e4047" (UID: "3fdb2e02-8713-4205-ae31-0af33f8e4047"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.872585 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3fdb2e02-8713-4205-ae31-0af33f8e4047" (UID: "3fdb2e02-8713-4205-ae31-0af33f8e4047"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.948025 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4jh4\" (UniqueName: \"kubernetes.io/projected/3fdb2e02-8713-4205-ae31-0af33f8e4047-kube-api-access-j4jh4\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.948353 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.948363 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:07 crc kubenswrapper[4948]: I0220 08:24:07.948372 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fdb2e02-8713-4205-ae31-0af33f8e4047-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.024691 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.024755 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.305543 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" event={"ID":"3fdb2e02-8713-4205-ae31-0af33f8e4047","Type":"ContainerDied","Data":"d20f6a9cd2d1756d07c023cf4fc3e418531e5c20161aa674031ec3aa6a8df0d2"} Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.305587 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d20f6a9cd2d1756d07c023cf4fc3e418531e5c20161aa674031ec3aa6a8df0d2" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.305659 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xt5t7" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.306707 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.306932 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.307173 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.307931 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.389775 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 20 08:24:08 crc kubenswrapper[4948]: E0220 08:24:08.390301 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fdb2e02-8713-4205-ae31-0af33f8e4047" containerName="nova-cell0-conductor-db-sync" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.390325 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fdb2e02-8713-4205-ae31-0af33f8e4047" containerName="nova-cell0-conductor-db-sync" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.390587 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fdb2e02-8713-4205-ae31-0af33f8e4047" containerName="nova-cell0-conductor-db-sync" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.391467 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.396920 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2tb2s" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.397379 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.400557 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.458432 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/873a6553-9637-4f2a-a743-f33cde7fc883-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.458534 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/873a6553-9637-4f2a-a743-f33cde7fc883-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.458641 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfd9z\" (UniqueName: \"kubernetes.io/projected/873a6553-9637-4f2a-a743-f33cde7fc883-kube-api-access-rfd9z\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.560382 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/873a6553-9637-4f2a-a743-f33cde7fc883-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.560447 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfd9z\" (UniqueName: \"kubernetes.io/projected/873a6553-9637-4f2a-a743-f33cde7fc883-kube-api-access-rfd9z\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.560516 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/873a6553-9637-4f2a-a743-f33cde7fc883-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.567020 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/873a6553-9637-4f2a-a743-f33cde7fc883-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.571150 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/873a6553-9637-4f2a-a743-f33cde7fc883-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.581415 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfd9z\" (UniqueName: \"kubernetes.io/projected/873a6553-9637-4f2a-a743-f33cde7fc883-kube-api-access-rfd9z\") pod \"nova-cell0-conductor-0\" (UID: \"873a6553-9637-4f2a-a743-f33cde7fc883\") " pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:08 crc kubenswrapper[4948]: I0220 08:24:08.730389 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:09 crc kubenswrapper[4948]: I0220 08:24:09.203657 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 20 08:24:09 crc kubenswrapper[4948]: I0220 08:24:09.316725 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"873a6553-9637-4f2a-a743-f33cde7fc883","Type":"ContainerStarted","Data":"d84333201b58aeae0039550c4336cd843c3f79fd281d4d0e31148878aaed4cf4"} Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.182317 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.231177 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.326918 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"873a6553-9637-4f2a-a743-f33cde7fc883","Type":"ContainerStarted","Data":"0ae670dd037318612ffa9a4d844e01a84b89b3a67099c7467316a31abc2e9825"} Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.327320 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.327333 4948 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.327905 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.338571 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.344675 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.34465484 podStartE2EDuration="2.34465484s" podCreationTimestamp="2026-02-20 08:24:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:10.342669141 +0000 UTC m=+1099.317163971" watchObservedRunningTime="2026-02-20 08:24:10.34465484 +0000 UTC m=+1099.319149670" Feb 20 08:24:10 crc kubenswrapper[4948]: I0220 08:24:10.384456 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 20 08:24:13 crc kubenswrapper[4948]: I0220 08:24:13.339304 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Feb 20 08:24:18 crc kubenswrapper[4948]: I0220 08:24:18.757236 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.356206 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-qnlrn"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.358850 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.363743 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.376338 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qnlrn"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.385389 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.490195 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.491281 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.498759 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.504323 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-scripts\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.504419 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j4xb\" (UniqueName: \"kubernetes.io/projected/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-kube-api-access-2j4xb\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.504507 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-config-data\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.504540 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.507437 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.581584 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.591537 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.594241 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.598109 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.606129 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-config-data\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.606193 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j4xb\" (UniqueName: \"kubernetes.io/projected/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-kube-api-access-2j4xb\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.606265 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-config-data\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.606283 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.606315 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.606358 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-scripts\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.606388 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl2cc\" (UniqueName: \"kubernetes.io/projected/60006453-1770-44da-bd53-89701da2fa43-kube-api-access-dl2cc\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.617898 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.620864 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-config-data\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.621546 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.626492 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.640070 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.644583 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-scripts\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.646398 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j4xb\" (UniqueName: \"kubernetes.io/projected/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-kube-api-access-2j4xb\") pod \"nova-cell0-cell-mapping-qnlrn\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.701708 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709333 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709384 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl2cc\" (UniqueName: \"kubernetes.io/projected/60006453-1770-44da-bd53-89701da2fa43-kube-api-access-dl2cc\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709441 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blbc4\" (UniqueName: \"kubernetes.io/projected/025de664-c84b-4bb4-b953-13d95e9eee5d-kube-api-access-blbc4\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709461 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-config-data\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709501 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ll5q\" (UniqueName: \"kubernetes.io/projected/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-kube-api-access-2ll5q\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709521 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-config-data\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709569 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709600 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025de664-c84b-4bb4-b953-13d95e9eee5d-logs\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709629 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.709649 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.715802 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.732553 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-config-data\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.739680 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.748681 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl2cc\" (UniqueName: \"kubernetes.io/projected/60006453-1770-44da-bd53-89701da2fa43-kube-api-access-dl2cc\") pod \"nova-scheduler-0\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.789212 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.795691 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.796009 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.799391 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.814289 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.814346 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.814395 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.814468 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blbc4\" (UniqueName: \"kubernetes.io/projected/025de664-c84b-4bb4-b953-13d95e9eee5d-kube-api-access-blbc4\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.814533 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ll5q\" (UniqueName: \"kubernetes.io/projected/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-kube-api-access-2ll5q\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.814563 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-config-data\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.814640 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025de664-c84b-4bb4-b953-13d95e9eee5d-logs\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.815198 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025de664-c84b-4bb4-b953-13d95e9eee5d-logs\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.824869 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.833891 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.834997 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.836699 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-config-data\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.841155 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.865889 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ll5q\" (UniqueName: \"kubernetes.io/projected/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-kube-api-access-2ll5q\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.869653 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blbc4\" (UniqueName: \"kubernetes.io/projected/025de664-c84b-4bb4-b953-13d95e9eee5d-kube-api-access-blbc4\") pod \"nova-api-0\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " pod="openstack/nova-api-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.911034 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x22px"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.912604 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.915999 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-config-data\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.916092 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxxm8\" (UniqueName: \"kubernetes.io/projected/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-kube-api-access-vxxm8\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.916558 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.916611 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-logs\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.943519 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x22px"] Feb 20 08:24:19 crc kubenswrapper[4948]: I0220 08:24:19.948659 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.018873 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-config\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.018920 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-svc\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.018960 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.019017 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxxm8\" (UniqueName: \"kubernetes.io/projected/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-kube-api-access-vxxm8\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.019074 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.019110 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh28w\" (UniqueName: \"kubernetes.io/projected/538c1b6f-424d-48e3-a92b-7534fbf865fc-kube-api-access-hh28w\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.019133 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.019175 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.019208 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-logs\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.019245 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-config-data\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.020002 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-logs\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.024732 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.025299 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-config-data\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.040000 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxxm8\" (UniqueName: \"kubernetes.io/projected/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-kube-api-access-vxxm8\") pod \"nova-metadata-0\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.121404 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.121758 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh28w\" (UniqueName: \"kubernetes.io/projected/538c1b6f-424d-48e3-a92b-7534fbf865fc-kube-api-access-hh28w\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.121950 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.122178 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-config\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.122258 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.122519 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-svc\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.122556 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.122913 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-config\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.122919 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.123460 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-svc\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.125418 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.139454 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh28w\" (UniqueName: \"kubernetes.io/projected/538c1b6f-424d-48e3-a92b-7534fbf865fc-kube-api-access-hh28w\") pod \"dnsmasq-dns-757b4f8459-x22px\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.158644 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.205185 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.259383 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.412072 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qnlrn"] Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.475878 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qnlrn" event={"ID":"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b","Type":"ContainerStarted","Data":"12945171d0e3aefdd46fee90ddbf80a6cc5d07d1053f4e8bc8f96de548e35fcc"} Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.493111 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.586823 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.616000 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jl8ck"] Feb 20 08:24:20 crc kubenswrapper[4948]: W0220 08:24:20.616669 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eea9d32_4e2a_4ad9_a402_7d7a483a3dff.slice/crio-4ce5c5ab612acce40e4a349276cab71f6a92423acc9a5a2deff1b1f12271b01e WatchSource:0}: Error finding container 4ce5c5ab612acce40e4a349276cab71f6a92423acc9a5a2deff1b1f12271b01e: Status 404 returned error can't find the container with id 4ce5c5ab612acce40e4a349276cab71f6a92423acc9a5a2deff1b1f12271b01e Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.617667 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.619883 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.620053 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.635685 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.646163 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jl8ck"] Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.674752 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.747358 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-config-data\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.747433 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-scripts\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.747553 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.747619 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czd6x\" (UniqueName: \"kubernetes.io/projected/93f8e582-bd6a-44d4-a203-ba9950efcada-kube-api-access-czd6x\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.850241 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-config-data\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.850305 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-scripts\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.850326 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.850350 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czd6x\" (UniqueName: \"kubernetes.io/projected/93f8e582-bd6a-44d4-a203-ba9950efcada-kube-api-access-czd6x\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.856562 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-config-data\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.857316 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-scripts\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.863460 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.865285 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czd6x\" (UniqueName: \"kubernetes.io/projected/93f8e582-bd6a-44d4-a203-ba9950efcada-kube-api-access-czd6x\") pod \"nova-cell1-conductor-db-sync-jl8ck\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:20 crc kubenswrapper[4948]: I0220 08:24:20.929267 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x22px"] Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.053715 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.488062 4948 generic.go:334] "Generic (PLEG): container finished" podID="538c1b6f-424d-48e3-a92b-7534fbf865fc" containerID="70074ef7c3184ccb90626f2f932107dda3ac2287a8b654203232efa428fd6294" exitCode=0 Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.488169 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-x22px" event={"ID":"538c1b6f-424d-48e3-a92b-7534fbf865fc","Type":"ContainerDied","Data":"70074ef7c3184ccb90626f2f932107dda3ac2287a8b654203232efa428fd6294"} Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.488405 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-x22px" event={"ID":"538c1b6f-424d-48e3-a92b-7534fbf865fc","Type":"ContainerStarted","Data":"ca84f60c8f59ce30814af7db2ac25fea0b071bd0c109ef4bdad869f451268b2b"} Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.491794 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ad615fe7-b01a-4e4c-85b1-e028e5cdd866","Type":"ContainerStarted","Data":"2e2fb5966e87b5f5075085440ad9aaba0868d3b137c6af57f293094a97faf543"} Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.500252 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff","Type":"ContainerStarted","Data":"4ce5c5ab612acce40e4a349276cab71f6a92423acc9a5a2deff1b1f12271b01e"} Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.503182 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"025de664-c84b-4bb4-b953-13d95e9eee5d","Type":"ContainerStarted","Data":"a2ceacf4091ad26a00570ff4d1bfb8fcf3b35d5c7789387f917c60bdba3eb6b8"} Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.515998 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"60006453-1770-44da-bd53-89701da2fa43","Type":"ContainerStarted","Data":"22ac2dae6a489cbbc45577f9a64760c4218b1a60e308183ad1c3246e8923567d"} Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.520552 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qnlrn" event={"ID":"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b","Type":"ContainerStarted","Data":"9a9d495cbe6ddcf0246661f3edacc2b8afe21e8ef43eca82e9140a67f21f1c6d"} Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.536389 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-qnlrn" podStartSLOduration=2.53637047 podStartE2EDuration="2.53637047s" podCreationTimestamp="2026-02-20 08:24:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:21.53596579 +0000 UTC m=+1110.510460600" watchObservedRunningTime="2026-02-20 08:24:21.53637047 +0000 UTC m=+1110.510865290" Feb 20 08:24:21 crc kubenswrapper[4948]: I0220 08:24:21.557808 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jl8ck"] Feb 20 08:24:21 crc kubenswrapper[4948]: W0220 08:24:21.562602 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93f8e582_bd6a_44d4_a203_ba9950efcada.slice/crio-bc36e2f68570b6936d5e5dbd27c44aba36cbc45ceca2d3d6a770bf8b08d1d326 WatchSource:0}: Error finding container bc36e2f68570b6936d5e5dbd27c44aba36cbc45ceca2d3d6a770bf8b08d1d326: Status 404 returned error can't find the container with id bc36e2f68570b6936d5e5dbd27c44aba36cbc45ceca2d3d6a770bf8b08d1d326 Feb 20 08:24:22 crc kubenswrapper[4948]: I0220 08:24:22.529632 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" event={"ID":"93f8e582-bd6a-44d4-a203-ba9950efcada","Type":"ContainerStarted","Data":"67cb9d3659cf07e0efb010d75eb62a0dddad421b940140c438f0a0714676a1ed"} Feb 20 08:24:22 crc kubenswrapper[4948]: I0220 08:24:22.529920 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" event={"ID":"93f8e582-bd6a-44d4-a203-ba9950efcada","Type":"ContainerStarted","Data":"bc36e2f68570b6936d5e5dbd27c44aba36cbc45ceca2d3d6a770bf8b08d1d326"} Feb 20 08:24:22 crc kubenswrapper[4948]: I0220 08:24:22.533651 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-x22px" event={"ID":"538c1b6f-424d-48e3-a92b-7534fbf865fc","Type":"ContainerStarted","Data":"3a9357bad2b3dca80dc83315104b35d29a1b980c7594ca0611bf4b94b36eb262"} Feb 20 08:24:22 crc kubenswrapper[4948]: I0220 08:24:22.548237 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" podStartSLOduration=2.548220019 podStartE2EDuration="2.548220019s" podCreationTimestamp="2026-02-20 08:24:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:22.545655567 +0000 UTC m=+1111.520150397" watchObservedRunningTime="2026-02-20 08:24:22.548220019 +0000 UTC m=+1111.522714829" Feb 20 08:24:22 crc kubenswrapper[4948]: I0220 08:24:22.595543 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-x22px" podStartSLOduration=3.5955207 podStartE2EDuration="3.5955207s" podCreationTimestamp="2026-02-20 08:24:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:22.578844838 +0000 UTC m=+1111.553339658" watchObservedRunningTime="2026-02-20 08:24:22.5955207 +0000 UTC m=+1111.570015520" Feb 20 08:24:23 crc kubenswrapper[4948]: I0220 08:24:23.481022 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:23 crc kubenswrapper[4948]: I0220 08:24:23.493648 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:23 crc kubenswrapper[4948]: I0220 08:24:23.541271 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.479882 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.534308 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-run-httpd\") pod \"0894350a-0df8-4c04-b7c7-fd8f93139959\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.534348 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-combined-ca-bundle\") pod \"0894350a-0df8-4c04-b7c7-fd8f93139959\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.534394 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txt8k\" (UniqueName: \"kubernetes.io/projected/0894350a-0df8-4c04-b7c7-fd8f93139959-kube-api-access-txt8k\") pod \"0894350a-0df8-4c04-b7c7-fd8f93139959\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.534416 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-scripts\") pod \"0894350a-0df8-4c04-b7c7-fd8f93139959\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.534453 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-config-data\") pod \"0894350a-0df8-4c04-b7c7-fd8f93139959\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.534473 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-sg-core-conf-yaml\") pod \"0894350a-0df8-4c04-b7c7-fd8f93139959\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.534581 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-log-httpd\") pod \"0894350a-0df8-4c04-b7c7-fd8f93139959\" (UID: \"0894350a-0df8-4c04-b7c7-fd8f93139959\") " Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.535252 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0894350a-0df8-4c04-b7c7-fd8f93139959" (UID: "0894350a-0df8-4c04-b7c7-fd8f93139959"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.535489 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0894350a-0df8-4c04-b7c7-fd8f93139959" (UID: "0894350a-0df8-4c04-b7c7-fd8f93139959"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.540242 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0894350a-0df8-4c04-b7c7-fd8f93139959-kube-api-access-txt8k" (OuterVolumeSpecName: "kube-api-access-txt8k") pod "0894350a-0df8-4c04-b7c7-fd8f93139959" (UID: "0894350a-0df8-4c04-b7c7-fd8f93139959"). InnerVolumeSpecName "kube-api-access-txt8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.541544 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-scripts" (OuterVolumeSpecName: "scripts") pod "0894350a-0df8-4c04-b7c7-fd8f93139959" (UID: "0894350a-0df8-4c04-b7c7-fd8f93139959"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.552135 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"025de664-c84b-4bb4-b953-13d95e9eee5d","Type":"ContainerStarted","Data":"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7"} Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.552176 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"025de664-c84b-4bb4-b953-13d95e9eee5d","Type":"ContainerStarted","Data":"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56"} Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.553528 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"60006453-1770-44da-bd53-89701da2fa43","Type":"ContainerStarted","Data":"789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a"} Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.558477 4948 generic.go:334] "Generic (PLEG): container finished" podID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerID="c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5" exitCode=137 Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.558540 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerDied","Data":"c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5"} Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.558561 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0894350a-0df8-4c04-b7c7-fd8f93139959","Type":"ContainerDied","Data":"6657eaf082a861ad8f0f29c4c00aaf35d258d94004b82734a54c5722b14133f5"} Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.558581 4948 scope.go:117] "RemoveContainer" containerID="c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.558700 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.565607 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ad615fe7-b01a-4e4c-85b1-e028e5cdd866","Type":"ContainerStarted","Data":"2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9"} Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.565721 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="ad615fe7-b01a-4e4c-85b1-e028e5cdd866" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9" gracePeriod=30 Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.576401 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0894350a-0df8-4c04-b7c7-fd8f93139959" (UID: "0894350a-0df8-4c04-b7c7-fd8f93139959"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.576576 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerName="nova-metadata-log" containerID="cri-o://b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a" gracePeriod=30 Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.576734 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff","Type":"ContainerStarted","Data":"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911"} Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.576773 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff","Type":"ContainerStarted","Data":"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a"} Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.576848 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerName="nova-metadata-metadata" containerID="cri-o://3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911" gracePeriod=30 Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.581777 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.580605578 podStartE2EDuration="5.581764532s" podCreationTimestamp="2026-02-20 08:24:19 +0000 UTC" firstStartedPulling="2026-02-20 08:24:20.590395671 +0000 UTC m=+1109.564907552" lastFinishedPulling="2026-02-20 08:24:23.591571686 +0000 UTC m=+1112.566066506" observedRunningTime="2026-02-20 08:24:24.572081108 +0000 UTC m=+1113.546575928" watchObservedRunningTime="2026-02-20 08:24:24.581764532 +0000 UTC m=+1113.556259342" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.593282 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.69013333 podStartE2EDuration="5.593262209s" podCreationTimestamp="2026-02-20 08:24:19 +0000 UTC" firstStartedPulling="2026-02-20 08:24:20.688467548 +0000 UTC m=+1109.662962368" lastFinishedPulling="2026-02-20 08:24:23.591596427 +0000 UTC m=+1112.566091247" observedRunningTime="2026-02-20 08:24:24.583709209 +0000 UTC m=+1113.558204019" watchObservedRunningTime="2026-02-20 08:24:24.593262209 +0000 UTC m=+1113.567757029" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.595804 4948 scope.go:117] "RemoveContainer" containerID="d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.624736 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.55623708 podStartE2EDuration="5.624716299s" podCreationTimestamp="2026-02-20 08:24:19 +0000 UTC" firstStartedPulling="2026-02-20 08:24:20.525205938 +0000 UTC m=+1109.499700758" lastFinishedPulling="2026-02-20 08:24:23.593685157 +0000 UTC m=+1112.568179977" observedRunningTime="2026-02-20 08:24:24.622475454 +0000 UTC m=+1113.596970274" watchObservedRunningTime="2026-02-20 08:24:24.624716299 +0000 UTC m=+1113.599211119" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.639608 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txt8k\" (UniqueName: \"kubernetes.io/projected/0894350a-0df8-4c04-b7c7-fd8f93139959-kube-api-access-txt8k\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.639948 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.639959 4948 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.640008 4948 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.640021 4948 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0894350a-0df8-4c04-b7c7-fd8f93139959-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.643250 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.6689121289999997 podStartE2EDuration="5.643240796s" podCreationTimestamp="2026-02-20 08:24:19 +0000 UTC" firstStartedPulling="2026-02-20 08:24:20.62182018 +0000 UTC m=+1109.596314990" lastFinishedPulling="2026-02-20 08:24:23.596148837 +0000 UTC m=+1112.570643657" observedRunningTime="2026-02-20 08:24:24.642225641 +0000 UTC m=+1113.616720461" watchObservedRunningTime="2026-02-20 08:24:24.643240796 +0000 UTC m=+1113.617735616" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.656924 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0894350a-0df8-4c04-b7c7-fd8f93139959" (UID: "0894350a-0df8-4c04-b7c7-fd8f93139959"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.691227 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-config-data" (OuterVolumeSpecName: "config-data") pod "0894350a-0df8-4c04-b7c7-fd8f93139959" (UID: "0894350a-0df8-4c04-b7c7-fd8f93139959"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.741250 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.741285 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0894350a-0df8-4c04-b7c7-fd8f93139959-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.845454 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.886826 4948 scope.go:117] "RemoveContainer" containerID="8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.906848 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.922322 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.940363 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:24 crc kubenswrapper[4948]: E0220 08:24:24.940706 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="ceilometer-notification-agent" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.940718 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="ceilometer-notification-agent" Feb 20 08:24:24 crc kubenswrapper[4948]: E0220 08:24:24.940743 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="ceilometer-central-agent" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.940750 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="ceilometer-central-agent" Feb 20 08:24:24 crc kubenswrapper[4948]: E0220 08:24:24.940760 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="sg-core" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.940766 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="sg-core" Feb 20 08:24:24 crc kubenswrapper[4948]: E0220 08:24:24.940779 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="proxy-httpd" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.940784 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="proxy-httpd" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.943774 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="ceilometer-central-agent" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.943800 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="ceilometer-notification-agent" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.943813 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="sg-core" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.943835 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" containerName="proxy-httpd" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.958321 4948 scope.go:117] "RemoveContainer" containerID="5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.959540 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.966946 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.967724 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 08:24:24 crc kubenswrapper[4948]: I0220 08:24:24.986488 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.019397 4948 scope.go:117] "RemoveContainer" containerID="c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5" Feb 20 08:24:25 crc kubenswrapper[4948]: E0220 08:24:25.022613 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5\": container with ID starting with c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5 not found: ID does not exist" containerID="c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.022658 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5"} err="failed to get container status \"c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5\": rpc error: code = NotFound desc = could not find container \"c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5\": container with ID starting with c4e7730b1940fa1741db63557d39a1291cbd6edeee74b043d33b1623adece0d5 not found: ID does not exist" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.022682 4948 scope.go:117] "RemoveContainer" containerID="d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed" Feb 20 08:24:25 crc kubenswrapper[4948]: E0220 08:24:25.026100 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed\": container with ID starting with d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed not found: ID does not exist" containerID="d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.026141 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed"} err="failed to get container status \"d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed\": rpc error: code = NotFound desc = could not find container \"d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed\": container with ID starting with d2ec8efadba57da267b73b627d540acadced09eb54d33e0b4318c801f4b817ed not found: ID does not exist" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.026168 4948 scope.go:117] "RemoveContainer" containerID="8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc" Feb 20 08:24:25 crc kubenswrapper[4948]: E0220 08:24:25.027659 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc\": container with ID starting with 8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc not found: ID does not exist" containerID="8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.027683 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc"} err="failed to get container status \"8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc\": rpc error: code = NotFound desc = could not find container \"8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc\": container with ID starting with 8372790439c71a0ea7e4da810ca880dff1b44a6d3fa022b0a9cbbafbf8e763cc not found: ID does not exist" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.027712 4948 scope.go:117] "RemoveContainer" containerID="5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4" Feb 20 08:24:25 crc kubenswrapper[4948]: E0220 08:24:25.044169 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4\": container with ID starting with 5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4 not found: ID does not exist" containerID="5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.044215 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4"} err="failed to get container status \"5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4\": rpc error: code = NotFound desc = could not find container \"5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4\": container with ID starting with 5a880cf140d0830077bc8b3d8a9a05a1bd68a08c91042ee55e74a3422f78c6e4 not found: ID does not exist" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.063592 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-scripts\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.063647 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-run-httpd\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.063672 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-config-data\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.063704 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrzbr\" (UniqueName: \"kubernetes.io/projected/ee88c030-4657-480c-99a0-90ee4c2b247c-kube-api-access-rrzbr\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.063732 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-log-httpd\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.063764 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.063816 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.111469 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.159534 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.165426 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-combined-ca-bundle\") pod \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.165523 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-config-data\") pod \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.165603 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxxm8\" (UniqueName: \"kubernetes.io/projected/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-kube-api-access-vxxm8\") pod \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.165689 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-logs\") pod \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\" (UID: \"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff\") " Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.165882 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.165966 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-scripts\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.166000 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-run-httpd\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.166021 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-config-data\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.166046 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrzbr\" (UniqueName: \"kubernetes.io/projected/ee88c030-4657-480c-99a0-90ee4c2b247c-kube-api-access-rrzbr\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.166072 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-log-httpd\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.166099 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.167545 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-run-httpd\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.167571 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-log-httpd\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.167772 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-logs" (OuterVolumeSpecName: "logs") pod "1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" (UID: "1eea9d32-4e2a-4ad9-a402-7d7a483a3dff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.171230 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.171620 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-config-data\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.172166 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.172938 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-scripts\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.174190 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-kube-api-access-vxxm8" (OuterVolumeSpecName: "kube-api-access-vxxm8") pod "1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" (UID: "1eea9d32-4e2a-4ad9-a402-7d7a483a3dff"). InnerVolumeSpecName "kube-api-access-vxxm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.191611 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrzbr\" (UniqueName: \"kubernetes.io/projected/ee88c030-4657-480c-99a0-90ee4c2b247c-kube-api-access-rrzbr\") pod \"ceilometer-0\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.195790 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" (UID: "1eea9d32-4e2a-4ad9-a402-7d7a483a3dff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.201272 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-config-data" (OuterVolumeSpecName: "config-data") pod "1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" (UID: "1eea9d32-4e2a-4ad9-a402-7d7a483a3dff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.267935 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.267986 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.268001 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.268013 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxxm8\" (UniqueName: \"kubernetes.io/projected/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff-kube-api-access-vxxm8\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.323109 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.589553 4948 generic.go:334] "Generic (PLEG): container finished" podID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerID="3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911" exitCode=0 Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.589584 4948 generic.go:334] "Generic (PLEG): container finished" podID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerID="b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a" exitCode=143 Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.589617 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff","Type":"ContainerDied","Data":"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911"} Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.589641 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff","Type":"ContainerDied","Data":"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a"} Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.589652 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1eea9d32-4e2a-4ad9-a402-7d7a483a3dff","Type":"ContainerDied","Data":"4ce5c5ab612acce40e4a349276cab71f6a92423acc9a5a2deff1b1f12271b01e"} Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.589667 4948 scope.go:117] "RemoveContainer" containerID="3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.589747 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.699472 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.704551 4948 scope.go:117] "RemoveContainer" containerID="b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.711745 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.730307 4948 scope.go:117] "RemoveContainer" containerID="3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911" Feb 20 08:24:25 crc kubenswrapper[4948]: E0220 08:24:25.736782 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911\": container with ID starting with 3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911 not found: ID does not exist" containerID="3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.736836 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911"} err="failed to get container status \"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911\": rpc error: code = NotFound desc = could not find container \"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911\": container with ID starting with 3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911 not found: ID does not exist" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.736892 4948 scope.go:117] "RemoveContainer" containerID="b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.737735 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0894350a-0df8-4c04-b7c7-fd8f93139959" path="/var/lib/kubelet/pods/0894350a-0df8-4c04-b7c7-fd8f93139959/volumes" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.738456 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" path="/var/lib/kubelet/pods/1eea9d32-4e2a-4ad9-a402-7d7a483a3dff/volumes" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.738966 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:25 crc kubenswrapper[4948]: E0220 08:24:25.739281 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerName="nova-metadata-log" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.739295 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerName="nova-metadata-log" Feb 20 08:24:25 crc kubenswrapper[4948]: E0220 08:24:25.739326 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerName="nova-metadata-metadata" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.739332 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerName="nova-metadata-metadata" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.739499 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerName="nova-metadata-metadata" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.739520 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eea9d32-4e2a-4ad9-a402-7d7a483a3dff" containerName="nova-metadata-log" Feb 20 08:24:25 crc kubenswrapper[4948]: E0220 08:24:25.740204 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a\": container with ID starting with b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a not found: ID does not exist" containerID="b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.740252 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a"} err="failed to get container status \"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a\": rpc error: code = NotFound desc = could not find container \"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a\": container with ID starting with b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a not found: ID does not exist" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.740274 4948 scope.go:117] "RemoveContainer" containerID="3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.741068 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.743873 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911"} err="failed to get container status \"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911\": rpc error: code = NotFound desc = could not find container \"3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911\": container with ID starting with 3fed0c16062dd9c721f4ae6c85d0591c043edf6f3931754416d90380a4013911 not found: ID does not exist" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.743910 4948 scope.go:117] "RemoveContainer" containerID="b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.744280 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.744361 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.747326 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a"} err="failed to get container status \"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a\": rpc error: code = NotFound desc = could not find container \"b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a\": container with ID starting with b4a821001a7cb14dc5e5191789693548c98621dfc170f94320c721e7b6198d8a not found: ID does not exist" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.749826 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.785186 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-config-data\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.785277 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.785295 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e86920-0969-4e9f-9094-be4fcc5e6909-logs\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.785322 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85fgw\" (UniqueName: \"kubernetes.io/projected/49e86920-0969-4e9f-9094-be4fcc5e6909-kube-api-access-85fgw\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.785336 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.785380 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: W0220 08:24:25.798950 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee88c030_4657_480c_99a0_90ee4c2b247c.slice/crio-7a7f2ca4a85292e0cdfdd80686c701b75dbf3b7d4a7696caa3251ac73f926e4a WatchSource:0}: Error finding container 7a7f2ca4a85292e0cdfdd80686c701b75dbf3b7d4a7696caa3251ac73f926e4a: Status 404 returned error can't find the container with id 7a7f2ca4a85292e0cdfdd80686c701b75dbf3b7d4a7696caa3251ac73f926e4a Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.886784 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-config-data\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.886846 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.886870 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e86920-0969-4e9f-9094-be4fcc5e6909-logs\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.886918 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85fgw\" (UniqueName: \"kubernetes.io/projected/49e86920-0969-4e9f-9094-be4fcc5e6909-kube-api-access-85fgw\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.886987 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.888941 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e86920-0969-4e9f-9094-be4fcc5e6909-logs\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.891174 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.893151 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-config-data\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.900833 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:25 crc kubenswrapper[4948]: I0220 08:24:25.913498 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85fgw\" (UniqueName: \"kubernetes.io/projected/49e86920-0969-4e9f-9094-be4fcc5e6909-kube-api-access-85fgw\") pod \"nova-metadata-0\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " pod="openstack/nova-metadata-0" Feb 20 08:24:26 crc kubenswrapper[4948]: I0220 08:24:26.057781 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:26 crc kubenswrapper[4948]: W0220 08:24:26.510332 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49e86920_0969_4e9f_9094_be4fcc5e6909.slice/crio-74fd9bc2eba68f1449066340009ee4df0ea7d5ad9b1c182e3bbc30ecc61defa2 WatchSource:0}: Error finding container 74fd9bc2eba68f1449066340009ee4df0ea7d5ad9b1c182e3bbc30ecc61defa2: Status 404 returned error can't find the container with id 74fd9bc2eba68f1449066340009ee4df0ea7d5ad9b1c182e3bbc30ecc61defa2 Feb 20 08:24:26 crc kubenswrapper[4948]: I0220 08:24:26.511380 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:26 crc kubenswrapper[4948]: I0220 08:24:26.601163 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49e86920-0969-4e9f-9094-be4fcc5e6909","Type":"ContainerStarted","Data":"74fd9bc2eba68f1449066340009ee4df0ea7d5ad9b1c182e3bbc30ecc61defa2"} Feb 20 08:24:26 crc kubenswrapper[4948]: I0220 08:24:26.602677 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerStarted","Data":"7a7f2ca4a85292e0cdfdd80686c701b75dbf3b7d4a7696caa3251ac73f926e4a"} Feb 20 08:24:27 crc kubenswrapper[4948]: I0220 08:24:27.609752 4948 generic.go:334] "Generic (PLEG): container finished" podID="21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" containerID="9a9d495cbe6ddcf0246661f3edacc2b8afe21e8ef43eca82e9140a67f21f1c6d" exitCode=0 Feb 20 08:24:27 crc kubenswrapper[4948]: I0220 08:24:27.609986 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qnlrn" event={"ID":"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b","Type":"ContainerDied","Data":"9a9d495cbe6ddcf0246661f3edacc2b8afe21e8ef43eca82e9140a67f21f1c6d"} Feb 20 08:24:27 crc kubenswrapper[4948]: I0220 08:24:27.615712 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49e86920-0969-4e9f-9094-be4fcc5e6909","Type":"ContainerStarted","Data":"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249"} Feb 20 08:24:27 crc kubenswrapper[4948]: I0220 08:24:27.615826 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49e86920-0969-4e9f-9094-be4fcc5e6909","Type":"ContainerStarted","Data":"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705"} Feb 20 08:24:27 crc kubenswrapper[4948]: I0220 08:24:27.658398 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.658377367 podStartE2EDuration="2.658377367s" podCreationTimestamp="2026-02-20 08:24:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:27.643721904 +0000 UTC m=+1116.618216724" watchObservedRunningTime="2026-02-20 08:24:27.658377367 +0000 UTC m=+1116.632872187" Feb 20 08:24:28 crc kubenswrapper[4948]: I0220 08:24:28.626794 4948 generic.go:334] "Generic (PLEG): container finished" podID="93f8e582-bd6a-44d4-a203-ba9950efcada" containerID="67cb9d3659cf07e0efb010d75eb62a0dddad421b940140c438f0a0714676a1ed" exitCode=0 Feb 20 08:24:28 crc kubenswrapper[4948]: I0220 08:24:28.626883 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" event={"ID":"93f8e582-bd6a-44d4-a203-ba9950efcada","Type":"ContainerDied","Data":"67cb9d3659cf07e0efb010d75eb62a0dddad421b940140c438f0a0714676a1ed"} Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.085641 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.186275 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-combined-ca-bundle\") pod \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.186412 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j4xb\" (UniqueName: \"kubernetes.io/projected/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-kube-api-access-2j4xb\") pod \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.186556 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-scripts\") pod \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.186603 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-config-data\") pod \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\" (UID: \"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b\") " Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.192390 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-scripts" (OuterVolumeSpecName: "scripts") pod "21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" (UID: "21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.193049 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-kube-api-access-2j4xb" (OuterVolumeSpecName: "kube-api-access-2j4xb") pod "21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" (UID: "21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b"). InnerVolumeSpecName "kube-api-access-2j4xb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.218159 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" (UID: "21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.234133 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-config-data" (OuterVolumeSpecName: "config-data") pod "21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" (UID: "21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.289359 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.290006 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.290048 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.290139 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j4xb\" (UniqueName: \"kubernetes.io/projected/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b-kube-api-access-2j4xb\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.641379 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qnlrn" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.641761 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qnlrn" event={"ID":"21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b","Type":"ContainerDied","Data":"12945171d0e3aefdd46fee90ddbf80a6cc5d07d1053f4e8bc8f96de548e35fcc"} Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.641857 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12945171d0e3aefdd46fee90ddbf80a6cc5d07d1053f4e8bc8f96de548e35fcc" Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.796401 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.796652 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerName="nova-api-log" containerID="cri-o://2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56" gracePeriod=30 Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.797118 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerName="nova-api-api" containerID="cri-o://a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7" gracePeriod=30 Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.816037 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.816252 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="60006453-1770-44da-bd53-89701da2fa43" containerName="nova-scheduler-scheduler" containerID="cri-o://789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a" gracePeriod=30 Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.826955 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.827177 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerName="nova-metadata-log" containerID="cri-o://f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705" gracePeriod=30 Feb 20 08:24:29 crc kubenswrapper[4948]: I0220 08:24:29.827310 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerName="nova-metadata-metadata" containerID="cri-o://32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249" gracePeriod=30 Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.163727 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.262168 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.314962 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-scripts\") pod \"93f8e582-bd6a-44d4-a203-ba9950efcada\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.315029 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-config-data\") pod \"93f8e582-bd6a-44d4-a203-ba9950efcada\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.315265 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czd6x\" (UniqueName: \"kubernetes.io/projected/93f8e582-bd6a-44d4-a203-ba9950efcada-kube-api-access-czd6x\") pod \"93f8e582-bd6a-44d4-a203-ba9950efcada\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.315296 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-combined-ca-bundle\") pod \"93f8e582-bd6a-44d4-a203-ba9950efcada\" (UID: \"93f8e582-bd6a-44d4-a203-ba9950efcada\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.324859 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93f8e582-bd6a-44d4-a203-ba9950efcada-kube-api-access-czd6x" (OuterVolumeSpecName: "kube-api-access-czd6x") pod "93f8e582-bd6a-44d4-a203-ba9950efcada" (UID: "93f8e582-bd6a-44d4-a203-ba9950efcada"). InnerVolumeSpecName "kube-api-access-czd6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.332102 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-scripts" (OuterVolumeSpecName: "scripts") pod "93f8e582-bd6a-44d4-a203-ba9950efcada" (UID: "93f8e582-bd6a-44d4-a203-ba9950efcada"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.378812 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5dgw7"] Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.379071 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" podUID="63124eef-9456-44d1-a8ba-cc8297af3e9b" containerName="dnsmasq-dns" containerID="cri-o://724a55a11670e45fe3d2ce7a82832ba87e341cd787eee57c186dd30931fb8402" gracePeriod=10 Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.395401 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-config-data" (OuterVolumeSpecName: "config-data") pod "93f8e582-bd6a-44d4-a203-ba9950efcada" (UID: "93f8e582-bd6a-44d4-a203-ba9950efcada"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.401054 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93f8e582-bd6a-44d4-a203-ba9950efcada" (UID: "93f8e582-bd6a-44d4-a203-ba9950efcada"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.418430 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.418455 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czd6x\" (UniqueName: \"kubernetes.io/projected/93f8e582-bd6a-44d4-a203-ba9950efcada-kube-api-access-czd6x\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.418472 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.418482 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93f8e582-bd6a-44d4-a203-ba9950efcada-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.544095 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.549209 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621409 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-combined-ca-bundle\") pod \"025de664-c84b-4bb4-b953-13d95e9eee5d\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621480 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-nova-metadata-tls-certs\") pod \"49e86920-0969-4e9f-9094-be4fcc5e6909\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621564 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85fgw\" (UniqueName: \"kubernetes.io/projected/49e86920-0969-4e9f-9094-be4fcc5e6909-kube-api-access-85fgw\") pod \"49e86920-0969-4e9f-9094-be4fcc5e6909\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621621 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-config-data\") pod \"49e86920-0969-4e9f-9094-be4fcc5e6909\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621743 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-config-data\") pod \"025de664-c84b-4bb4-b953-13d95e9eee5d\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621791 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e86920-0969-4e9f-9094-be4fcc5e6909-logs\") pod \"49e86920-0969-4e9f-9094-be4fcc5e6909\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621822 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025de664-c84b-4bb4-b953-13d95e9eee5d-logs\") pod \"025de664-c84b-4bb4-b953-13d95e9eee5d\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621869 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blbc4\" (UniqueName: \"kubernetes.io/projected/025de664-c84b-4bb4-b953-13d95e9eee5d-kube-api-access-blbc4\") pod \"025de664-c84b-4bb4-b953-13d95e9eee5d\" (UID: \"025de664-c84b-4bb4-b953-13d95e9eee5d\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.621920 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-combined-ca-bundle\") pod \"49e86920-0969-4e9f-9094-be4fcc5e6909\" (UID: \"49e86920-0969-4e9f-9094-be4fcc5e6909\") " Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.626413 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/025de664-c84b-4bb4-b953-13d95e9eee5d-logs" (OuterVolumeSpecName: "logs") pod "025de664-c84b-4bb4-b953-13d95e9eee5d" (UID: "025de664-c84b-4bb4-b953-13d95e9eee5d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.626724 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49e86920-0969-4e9f-9094-be4fcc5e6909-logs" (OuterVolumeSpecName: "logs") pod "49e86920-0969-4e9f-9094-be4fcc5e6909" (UID: "49e86920-0969-4e9f-9094-be4fcc5e6909"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.632560 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49e86920-0969-4e9f-9094-be4fcc5e6909-kube-api-access-85fgw" (OuterVolumeSpecName: "kube-api-access-85fgw") pod "49e86920-0969-4e9f-9094-be4fcc5e6909" (UID: "49e86920-0969-4e9f-9094-be4fcc5e6909"). InnerVolumeSpecName "kube-api-access-85fgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.642182 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/025de664-c84b-4bb4-b953-13d95e9eee5d-kube-api-access-blbc4" (OuterVolumeSpecName: "kube-api-access-blbc4") pod "025de664-c84b-4bb4-b953-13d95e9eee5d" (UID: "025de664-c84b-4bb4-b953-13d95e9eee5d"). InnerVolumeSpecName "kube-api-access-blbc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.651506 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerStarted","Data":"1630e56297ae1cd7ec343d1b51485df834ac4ef7ce2a756bf5a88dac19b42159"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.653507 4948 generic.go:334] "Generic (PLEG): container finished" podID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerID="a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7" exitCode=0 Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.653670 4948 generic.go:334] "Generic (PLEG): container finished" podID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerID="2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56" exitCode=143 Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.653571 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"025de664-c84b-4bb4-b953-13d95e9eee5d","Type":"ContainerDied","Data":"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.653571 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.654099 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"025de664-c84b-4bb4-b953-13d95e9eee5d","Type":"ContainerDied","Data":"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.654121 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"025de664-c84b-4bb4-b953-13d95e9eee5d","Type":"ContainerDied","Data":"a2ceacf4091ad26a00570ff4d1bfb8fcf3b35d5c7789387f917c60bdba3eb6b8"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.654138 4948 scope.go:117] "RemoveContainer" containerID="a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.666328 4948 generic.go:334] "Generic (PLEG): container finished" podID="63124eef-9456-44d1-a8ba-cc8297af3e9b" containerID="724a55a11670e45fe3d2ce7a82832ba87e341cd787eee57c186dd30931fb8402" exitCode=0 Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.666431 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" event={"ID":"63124eef-9456-44d1-a8ba-cc8297af3e9b","Type":"ContainerDied","Data":"724a55a11670e45fe3d2ce7a82832ba87e341cd787eee57c186dd30931fb8402"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.726687 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e86920-0969-4e9f-9094-be4fcc5e6909-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.726713 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025de664-c84b-4bb4-b953-13d95e9eee5d-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.726722 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blbc4\" (UniqueName: \"kubernetes.io/projected/025de664-c84b-4bb4-b953-13d95e9eee5d-kube-api-access-blbc4\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.726730 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85fgw\" (UniqueName: \"kubernetes.io/projected/49e86920-0969-4e9f-9094-be4fcc5e6909-kube-api-access-85fgw\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.742432 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "025de664-c84b-4bb4-b953-13d95e9eee5d" (UID: "025de664-c84b-4bb4-b953-13d95e9eee5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.742722 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" event={"ID":"93f8e582-bd6a-44d4-a203-ba9950efcada","Type":"ContainerDied","Data":"bc36e2f68570b6936d5e5dbd27c44aba36cbc45ceca2d3d6a770bf8b08d1d326"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.742750 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc36e2f68570b6936d5e5dbd27c44aba36cbc45ceca2d3d6a770bf8b08d1d326" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.742833 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jl8ck" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.768654 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-config-data" (OuterVolumeSpecName: "config-data") pod "49e86920-0969-4e9f-9094-be4fcc5e6909" (UID: "49e86920-0969-4e9f-9094-be4fcc5e6909"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.778412 4948 scope.go:117] "RemoveContainer" containerID="2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.801185 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "49e86920-0969-4e9f-9094-be4fcc5e6909" (UID: "49e86920-0969-4e9f-9094-be4fcc5e6909"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.805109 4948 generic.go:334] "Generic (PLEG): container finished" podID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerID="32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249" exitCode=0 Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.805140 4948 generic.go:334] "Generic (PLEG): container finished" podID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerID="f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705" exitCode=143 Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.805161 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49e86920-0969-4e9f-9094-be4fcc5e6909","Type":"ContainerDied","Data":"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.805192 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49e86920-0969-4e9f-9094-be4fcc5e6909","Type":"ContainerDied","Data":"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.805203 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49e86920-0969-4e9f-9094-be4fcc5e6909","Type":"ContainerDied","Data":"74fd9bc2eba68f1449066340009ee4df0ea7d5ad9b1c182e3bbc30ecc61defa2"} Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.805261 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.816335 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "49e86920-0969-4e9f-9094-be4fcc5e6909" (UID: "49e86920-0969-4e9f-9094-be4fcc5e6909"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.825681 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 20 08:24:30 crc kubenswrapper[4948]: E0220 08:24:30.826284 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93f8e582-bd6a-44d4-a203-ba9950efcada" containerName="nova-cell1-conductor-db-sync" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826296 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="93f8e582-bd6a-44d4-a203-ba9950efcada" containerName="nova-cell1-conductor-db-sync" Feb 20 08:24:30 crc kubenswrapper[4948]: E0220 08:24:30.826308 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" containerName="nova-manage" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826314 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" containerName="nova-manage" Feb 20 08:24:30 crc kubenswrapper[4948]: E0220 08:24:30.826332 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerName="nova-metadata-metadata" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826338 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerName="nova-metadata-metadata" Feb 20 08:24:30 crc kubenswrapper[4948]: E0220 08:24:30.826357 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerName="nova-api-api" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826363 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerName="nova-api-api" Feb 20 08:24:30 crc kubenswrapper[4948]: E0220 08:24:30.826377 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerName="nova-metadata-log" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826383 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerName="nova-metadata-log" Feb 20 08:24:30 crc kubenswrapper[4948]: E0220 08:24:30.826392 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerName="nova-api-log" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826398 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerName="nova-api-log" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826554 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" containerName="nova-manage" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826566 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerName="nova-metadata-metadata" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826579 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerName="nova-api-log" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826587 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" containerName="nova-api-api" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826595 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" containerName="nova-metadata-log" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.826604 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="93f8e582-bd6a-44d4-a203-ba9950efcada" containerName="nova-cell1-conductor-db-sync" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.827215 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.830337 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-config-data" (OuterVolumeSpecName: "config-data") pod "025de664-c84b-4bb4-b953-13d95e9eee5d" (UID: "025de664-c84b-4bb4-b953-13d95e9eee5d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.832174 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.832195 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.832207 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.832217 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/025de664-c84b-4bb4-b953-13d95e9eee5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.832228 4948 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/49e86920-0969-4e9f-9094-be4fcc5e6909-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.832423 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.834599 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.874233 4948 scope.go:117] "RemoveContainer" containerID="a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7" Feb 20 08:24:30 crc kubenswrapper[4948]: E0220 08:24:30.876394 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7\": container with ID starting with a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7 not found: ID does not exist" containerID="a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.876434 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7"} err="failed to get container status \"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7\": rpc error: code = NotFound desc = could not find container \"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7\": container with ID starting with a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7 not found: ID does not exist" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.876458 4948 scope.go:117] "RemoveContainer" containerID="2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56" Feb 20 08:24:30 crc kubenswrapper[4948]: E0220 08:24:30.876799 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56\": container with ID starting with 2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56 not found: ID does not exist" containerID="2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.876814 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56"} err="failed to get container status \"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56\": rpc error: code = NotFound desc = could not find container \"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56\": container with ID starting with 2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56 not found: ID does not exist" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.876825 4948 scope.go:117] "RemoveContainer" containerID="a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.877034 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7"} err="failed to get container status \"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7\": rpc error: code = NotFound desc = could not find container \"a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7\": container with ID starting with a97ca7b16ac93a0a6c0542fc8c2ef5448be6b3a3c55981c28fa0d60c2271d6a7 not found: ID does not exist" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.877047 4948 scope.go:117] "RemoveContainer" containerID="2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.877204 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56"} err="failed to get container status \"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56\": rpc error: code = NotFound desc = could not find container \"2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56\": container with ID starting with 2be4f046b391712a4b0aeb2b102aa5abe53d186c277e20fb17a13476ca768a56 not found: ID does not exist" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.877217 4948 scope.go:117] "RemoveContainer" containerID="32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.934105 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e801ada9-0247-4b29-b262-04637e1f8452-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.934234 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e801ada9-0247-4b29-b262-04637e1f8452-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.934284 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dl7b\" (UniqueName: \"kubernetes.io/projected/e801ada9-0247-4b29-b262-04637e1f8452-kube-api-access-2dl7b\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:30 crc kubenswrapper[4948]: I0220 08:24:30.948077 4948 scope.go:117] "RemoveContainer" containerID="f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.024496 4948 scope.go:117] "RemoveContainer" containerID="32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249" Feb 20 08:24:31 crc kubenswrapper[4948]: E0220 08:24:31.025456 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249\": container with ID starting with 32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249 not found: ID does not exist" containerID="32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.025497 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249"} err="failed to get container status \"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249\": rpc error: code = NotFound desc = could not find container \"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249\": container with ID starting with 32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249 not found: ID does not exist" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.025525 4948 scope.go:117] "RemoveContainer" containerID="f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705" Feb 20 08:24:31 crc kubenswrapper[4948]: E0220 08:24:31.025955 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705\": container with ID starting with f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705 not found: ID does not exist" containerID="f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.025993 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705"} err="failed to get container status \"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705\": rpc error: code = NotFound desc = could not find container \"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705\": container with ID starting with f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705 not found: ID does not exist" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.026006 4948 scope.go:117] "RemoveContainer" containerID="32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.026814 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249"} err="failed to get container status \"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249\": rpc error: code = NotFound desc = could not find container \"32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249\": container with ID starting with 32ca44aabcc30785b2b30640adb7c4210d75a93230b16924e858acb20c880249 not found: ID does not exist" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.026835 4948 scope.go:117] "RemoveContainer" containerID="f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.027016 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705"} err="failed to get container status \"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705\": rpc error: code = NotFound desc = could not find container \"f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705\": container with ID starting with f17de108ac21a152d3ce593e80f1edf60f3323977e65e8edc22919cbf9196705 not found: ID does not exist" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.035992 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dl7b\" (UniqueName: \"kubernetes.io/projected/e801ada9-0247-4b29-b262-04637e1f8452-kube-api-access-2dl7b\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.036475 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e801ada9-0247-4b29-b262-04637e1f8452-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.036554 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e801ada9-0247-4b29-b262-04637e1f8452-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.042207 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e801ada9-0247-4b29-b262-04637e1f8452-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.055458 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.059715 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e801ada9-0247-4b29-b262-04637e1f8452-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.063957 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.066050 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dl7b\" (UniqueName: \"kubernetes.io/projected/e801ada9-0247-4b29-b262-04637e1f8452-kube-api-access-2dl7b\") pod \"nova-cell1-conductor-0\" (UID: \"e801ada9-0247-4b29-b262-04637e1f8452\") " pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.072510 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.074099 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.076495 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.083684 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.136884 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.138722 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b950a35-d978-4338-88a6-ec96c9af048e-logs\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.138890 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.139220 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-config-data\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.139281 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96ggh\" (UniqueName: \"kubernetes.io/projected/4b950a35-d978-4338-88a6-ec96c9af048e-kube-api-access-96ggh\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.151485 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.163208 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.177328 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: E0220 08:24:31.177805 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63124eef-9456-44d1-a8ba-cc8297af3e9b" containerName="dnsmasq-dns" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.177826 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="63124eef-9456-44d1-a8ba-cc8297af3e9b" containerName="dnsmasq-dns" Feb 20 08:24:31 crc kubenswrapper[4948]: E0220 08:24:31.177839 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63124eef-9456-44d1-a8ba-cc8297af3e9b" containerName="init" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.177847 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="63124eef-9456-44d1-a8ba-cc8297af3e9b" containerName="init" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.178104 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="63124eef-9456-44d1-a8ba-cc8297af3e9b" containerName="dnsmasq-dns" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.179118 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.181649 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.181876 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.199673 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.241390 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-sb\") pod \"63124eef-9456-44d1-a8ba-cc8297af3e9b\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.241781 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-nb\") pod \"63124eef-9456-44d1-a8ba-cc8297af3e9b\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.241812 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2cgj\" (UniqueName: \"kubernetes.io/projected/63124eef-9456-44d1-a8ba-cc8297af3e9b-kube-api-access-l2cgj\") pod \"63124eef-9456-44d1-a8ba-cc8297af3e9b\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.241866 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-swift-storage-0\") pod \"63124eef-9456-44d1-a8ba-cc8297af3e9b\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.241938 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-config\") pod \"63124eef-9456-44d1-a8ba-cc8297af3e9b\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.242043 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-svc\") pod \"63124eef-9456-44d1-a8ba-cc8297af3e9b\" (UID: \"63124eef-9456-44d1-a8ba-cc8297af3e9b\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.242326 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b950a35-d978-4338-88a6-ec96c9af048e-logs\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.242351 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7ch5\" (UniqueName: \"kubernetes.io/projected/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-kube-api-access-p7ch5\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.242398 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.242416 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.242435 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-config-data\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.242822 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b950a35-d978-4338-88a6-ec96c9af048e-logs\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.243880 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.244462 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.244492 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-logs\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.244553 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-config-data\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.244572 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96ggh\" (UniqueName: \"kubernetes.io/projected/4b950a35-d978-4338-88a6-ec96c9af048e-kube-api-access-96ggh\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.252223 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63124eef-9456-44d1-a8ba-cc8297af3e9b-kube-api-access-l2cgj" (OuterVolumeSpecName: "kube-api-access-l2cgj") pod "63124eef-9456-44d1-a8ba-cc8297af3e9b" (UID: "63124eef-9456-44d1-a8ba-cc8297af3e9b"). InnerVolumeSpecName "kube-api-access-l2cgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.253808 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-config-data\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.262071 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.265283 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96ggh\" (UniqueName: \"kubernetes.io/projected/4b950a35-d978-4338-88a6-ec96c9af048e-kube-api-access-96ggh\") pod \"nova-api-0\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.317936 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-config" (OuterVolumeSpecName: "config") pod "63124eef-9456-44d1-a8ba-cc8297af3e9b" (UID: "63124eef-9456-44d1-a8ba-cc8297af3e9b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.317683 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "63124eef-9456-44d1-a8ba-cc8297af3e9b" (UID: "63124eef-9456-44d1-a8ba-cc8297af3e9b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.325316 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "63124eef-9456-44d1-a8ba-cc8297af3e9b" (UID: "63124eef-9456-44d1-a8ba-cc8297af3e9b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.345719 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "63124eef-9456-44d1-a8ba-cc8297af3e9b" (UID: "63124eef-9456-44d1-a8ba-cc8297af3e9b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.345974 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346032 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-logs\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346151 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7ch5\" (UniqueName: \"kubernetes.io/projected/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-kube-api-access-p7ch5\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346198 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346260 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-config-data\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346357 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "63124eef-9456-44d1-a8ba-cc8297af3e9b" (UID: "63124eef-9456-44d1-a8ba-cc8297af3e9b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346365 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346431 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2cgj\" (UniqueName: \"kubernetes.io/projected/63124eef-9456-44d1-a8ba-cc8297af3e9b-kube-api-access-l2cgj\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346494 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346506 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.346515 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.347387 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-logs\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.351915 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.354162 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.354402 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-config-data\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.364837 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7ch5\" (UniqueName: \"kubernetes.io/projected/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-kube-api-access-p7ch5\") pod \"nova-metadata-0\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.448531 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63124eef-9456-44d1-a8ba-cc8297af3e9b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.448708 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.524770 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.571407 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.653676 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl2cc\" (UniqueName: \"kubernetes.io/projected/60006453-1770-44da-bd53-89701da2fa43-kube-api-access-dl2cc\") pod \"60006453-1770-44da-bd53-89701da2fa43\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.653726 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-combined-ca-bundle\") pod \"60006453-1770-44da-bd53-89701da2fa43\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.653858 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-config-data\") pod \"60006453-1770-44da-bd53-89701da2fa43\" (UID: \"60006453-1770-44da-bd53-89701da2fa43\") " Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.660390 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60006453-1770-44da-bd53-89701da2fa43-kube-api-access-dl2cc" (OuterVolumeSpecName: "kube-api-access-dl2cc") pod "60006453-1770-44da-bd53-89701da2fa43" (UID: "60006453-1770-44da-bd53-89701da2fa43"). InnerVolumeSpecName "kube-api-access-dl2cc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.681497 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60006453-1770-44da-bd53-89701da2fa43" (UID: "60006453-1770-44da-bd53-89701da2fa43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.688976 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-config-data" (OuterVolumeSpecName: "config-data") pod "60006453-1770-44da-bd53-89701da2fa43" (UID: "60006453-1770-44da-bd53-89701da2fa43"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.761593 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl2cc\" (UniqueName: \"kubernetes.io/projected/60006453-1770-44da-bd53-89701da2fa43-kube-api-access-dl2cc\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.761897 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.761915 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60006453-1770-44da-bd53-89701da2fa43-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.768290 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="025de664-c84b-4bb4-b953-13d95e9eee5d" path="/var/lib/kubelet/pods/025de664-c84b-4bb4-b953-13d95e9eee5d/volumes" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.769453 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49e86920-0969-4e9f-9094-be4fcc5e6909" path="/var/lib/kubelet/pods/49e86920-0969-4e9f-9094-be4fcc5e6909/volumes" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.771595 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.844374 4948 generic.go:334] "Generic (PLEG): container finished" podID="60006453-1770-44da-bd53-89701da2fa43" containerID="789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a" exitCode=0 Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.844540 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.844638 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"60006453-1770-44da-bd53-89701da2fa43","Type":"ContainerDied","Data":"789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a"} Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.844697 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"60006453-1770-44da-bd53-89701da2fa43","Type":"ContainerDied","Data":"22ac2dae6a489cbbc45577f9a64760c4218b1a60e308183ad1c3246e8923567d"} Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.844747 4948 scope.go:117] "RemoveContainer" containerID="789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.855787 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerStarted","Data":"1bcad95c714fce375a1b0a71dbbd9424565d0d8570fd34397817043b9bfe2e88"} Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.858750 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" event={"ID":"63124eef-9456-44d1-a8ba-cc8297af3e9b","Type":"ContainerDied","Data":"b02b398d101ad6f54b2deb0825184d401d1e5b0ad4e38257f4120a442c5a5509"} Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.858840 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-5dgw7" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.860025 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e801ada9-0247-4b29-b262-04637e1f8452","Type":"ContainerStarted","Data":"4b659e360d8bb9cff850961f25c15821b1a120efd9472c53adb045fc0c3cbcca"} Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.877419 4948 scope.go:117] "RemoveContainer" containerID="789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a" Feb 20 08:24:31 crc kubenswrapper[4948]: E0220 08:24:31.878478 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a\": container with ID starting with 789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a not found: ID does not exist" containerID="789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.878554 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a"} err="failed to get container status \"789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a\": rpc error: code = NotFound desc = could not find container \"789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a\": container with ID starting with 789d33d79de6243b08689af1b9cf996bdaa6473c8e9a5246cde2871757156c6a not found: ID does not exist" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.878609 4948 scope.go:117] "RemoveContainer" containerID="724a55a11670e45fe3d2ce7a82832ba87e341cd787eee57c186dd30931fb8402" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.918676 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.931646 4948 scope.go:117] "RemoveContainer" containerID="56e628ea61223b6bdd0b21738d110aaaa403a55ddd583614f61bc57a6cdb16fd" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.940932 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.953815 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5dgw7"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.972327 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5dgw7"] Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.981039 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:31 crc kubenswrapper[4948]: E0220 08:24:31.981729 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60006453-1770-44da-bd53-89701da2fa43" containerName="nova-scheduler-scheduler" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.981766 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="60006453-1770-44da-bd53-89701da2fa43" containerName="nova-scheduler-scheduler" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.982060 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="60006453-1770-44da-bd53-89701da2fa43" containerName="nova-scheduler-scheduler" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.984522 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.987172 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 20 08:24:31 crc kubenswrapper[4948]: I0220 08:24:31.991480 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.031183 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:24:32 crc kubenswrapper[4948]: W0220 08:24:32.045800 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod659ea49b_7e8f_4c95_ac69_ec743d0f95ed.slice/crio-aaa4c4ee0564a915596564697d18943937054d1b42224076a88c2793a4afd7d0 WatchSource:0}: Error finding container aaa4c4ee0564a915596564697d18943937054d1b42224076a88c2793a4afd7d0: Status 404 returned error can't find the container with id aaa4c4ee0564a915596564697d18943937054d1b42224076a88c2793a4afd7d0 Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.069100 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-config-data\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.069567 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb4nn\" (UniqueName: \"kubernetes.io/projected/a76350cb-da45-403e-8934-ef53b257ea27-kube-api-access-cb4nn\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.069692 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: W0220 08:24:32.107152 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b950a35_d978_4338_88a6_ec96c9af048e.slice/crio-d66a33a2d7dbdb39f5f061898cb4a5149bc962e88f36c5fc69dae56ca6bdd17a WatchSource:0}: Error finding container d66a33a2d7dbdb39f5f061898cb4a5149bc962e88f36c5fc69dae56ca6bdd17a: Status 404 returned error can't find the container with id d66a33a2d7dbdb39f5f061898cb4a5149bc962e88f36c5fc69dae56ca6bdd17a Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.108360 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.171023 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-config-data\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.171056 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb4nn\" (UniqueName: \"kubernetes.io/projected/a76350cb-da45-403e-8934-ef53b257ea27-kube-api-access-cb4nn\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.171130 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.179025 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-config-data\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.179542 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.186863 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb4nn\" (UniqueName: \"kubernetes.io/projected/a76350cb-da45-403e-8934-ef53b257ea27-kube-api-access-cb4nn\") pod \"nova-scheduler-0\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.273094 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.695186 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:24:32 crc kubenswrapper[4948]: W0220 08:24:32.697709 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda76350cb_da45_403e_8934_ef53b257ea27.slice/crio-6fbbcc32e0efcd622515f0c849bc2e7cae40139cc47cc3aeb9187b65bfe28ca5 WatchSource:0}: Error finding container 6fbbcc32e0efcd622515f0c849bc2e7cae40139cc47cc3aeb9187b65bfe28ca5: Status 404 returned error can't find the container with id 6fbbcc32e0efcd622515f0c849bc2e7cae40139cc47cc3aeb9187b65bfe28ca5 Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.869121 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a76350cb-da45-403e-8934-ef53b257ea27","Type":"ContainerStarted","Data":"6fbbcc32e0efcd622515f0c849bc2e7cae40139cc47cc3aeb9187b65bfe28ca5"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.871849 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e801ada9-0247-4b29-b262-04637e1f8452","Type":"ContainerStarted","Data":"681c4bfe91487bec6fbff0de13c9517419b14290b69f54162d9b1bbc3572c765"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.872497 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.876868 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b950a35-d978-4338-88a6-ec96c9af048e","Type":"ContainerStarted","Data":"6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.876897 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b950a35-d978-4338-88a6-ec96c9af048e","Type":"ContainerStarted","Data":"d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.876908 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b950a35-d978-4338-88a6-ec96c9af048e","Type":"ContainerStarted","Data":"d66a33a2d7dbdb39f5f061898cb4a5149bc962e88f36c5fc69dae56ca6bdd17a"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.879399 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"659ea49b-7e8f-4c95-ac69-ec743d0f95ed","Type":"ContainerStarted","Data":"22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.879428 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"659ea49b-7e8f-4c95-ac69-ec743d0f95ed","Type":"ContainerStarted","Data":"63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.879441 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"659ea49b-7e8f-4c95-ac69-ec743d0f95ed","Type":"ContainerStarted","Data":"aaa4c4ee0564a915596564697d18943937054d1b42224076a88c2793a4afd7d0"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.884323 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerStarted","Data":"0574f1ab13d44d1f98f65e5fc2fb1727a44b35bb855f4f5336cf1da833d95b3a"} Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.899070 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.899046777 podStartE2EDuration="2.899046777s" podCreationTimestamp="2026-02-20 08:24:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:32.88760045 +0000 UTC m=+1121.862095270" watchObservedRunningTime="2026-02-20 08:24:32.899046777 +0000 UTC m=+1121.873541617" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.914698 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.914677324 podStartE2EDuration="1.914677324s" podCreationTimestamp="2026-02-20 08:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:32.908593937 +0000 UTC m=+1121.883088787" watchObservedRunningTime="2026-02-20 08:24:32.914677324 +0000 UTC m=+1121.889172154" Feb 20 08:24:32 crc kubenswrapper[4948]: I0220 08:24:32.929641 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.929622974 podStartE2EDuration="1.929622974s" podCreationTimestamp="2026-02-20 08:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:32.928411695 +0000 UTC m=+1121.902906535" watchObservedRunningTime="2026-02-20 08:24:32.929622974 +0000 UTC m=+1121.904117794" Feb 20 08:24:33 crc kubenswrapper[4948]: I0220 08:24:33.749814 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60006453-1770-44da-bd53-89701da2fa43" path="/var/lib/kubelet/pods/60006453-1770-44da-bd53-89701da2fa43/volumes" Feb 20 08:24:33 crc kubenswrapper[4948]: I0220 08:24:33.753170 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63124eef-9456-44d1-a8ba-cc8297af3e9b" path="/var/lib/kubelet/pods/63124eef-9456-44d1-a8ba-cc8297af3e9b/volumes" Feb 20 08:24:33 crc kubenswrapper[4948]: I0220 08:24:33.897114 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a76350cb-da45-403e-8934-ef53b257ea27","Type":"ContainerStarted","Data":"7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787"} Feb 20 08:24:33 crc kubenswrapper[4948]: I0220 08:24:33.931444 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.9314236 podStartE2EDuration="2.9314236s" podCreationTimestamp="2026-02-20 08:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:33.915587688 +0000 UTC m=+1122.890082508" watchObservedRunningTime="2026-02-20 08:24:33.9314236 +0000 UTC m=+1122.905918420" Feb 20 08:24:34 crc kubenswrapper[4948]: I0220 08:24:34.911754 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerStarted","Data":"430724be7dcee7dfd82861aa2d26928ec4bed184137610d057128b4eb3d23e14"} Feb 20 08:24:34 crc kubenswrapper[4948]: I0220 08:24:34.951847 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.677508246 podStartE2EDuration="10.951818044s" podCreationTimestamp="2026-02-20 08:24:24 +0000 UTC" firstStartedPulling="2026-02-20 08:24:25.800995295 +0000 UTC m=+1114.775490135" lastFinishedPulling="2026-02-20 08:24:34.075305103 +0000 UTC m=+1123.049799933" observedRunningTime="2026-02-20 08:24:34.941510935 +0000 UTC m=+1123.916005795" watchObservedRunningTime="2026-02-20 08:24:34.951818044 +0000 UTC m=+1123.926312904" Feb 20 08:24:35 crc kubenswrapper[4948]: I0220 08:24:35.926465 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 08:24:36 crc kubenswrapper[4948]: I0220 08:24:36.282966 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 20 08:24:36 crc kubenswrapper[4948]: I0220 08:24:36.525739 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 20 08:24:36 crc kubenswrapper[4948]: I0220 08:24:36.525848 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 20 08:24:37 crc kubenswrapper[4948]: I0220 08:24:37.273845 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.025362 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.025748 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.025804 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.026438 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8f0b8a806330c630468951158113ee0ecaad5dbddb81b09c063752a17f21282c"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.026490 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://8f0b8a806330c630468951158113ee0ecaad5dbddb81b09c063752a17f21282c" gracePeriod=600 Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.967676 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="8f0b8a806330c630468951158113ee0ecaad5dbddb81b09c063752a17f21282c" exitCode=0 Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.967802 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"8f0b8a806330c630468951158113ee0ecaad5dbddb81b09c063752a17f21282c"} Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.968226 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"d9efd6716995fcdcebdcf01e76fc7927735e34b1da4918444f90e59bd0f6d815"} Feb 20 08:24:38 crc kubenswrapper[4948]: I0220 08:24:38.968275 4948 scope.go:117] "RemoveContainer" containerID="df8e056c06ac241c46a927c4e3f5a28f39d042bf13a8c672499d528786dbbbb0" Feb 20 08:24:41 crc kubenswrapper[4948]: I0220 08:24:41.449787 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 08:24:41 crc kubenswrapper[4948]: I0220 08:24:41.450463 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 08:24:41 crc kubenswrapper[4948]: I0220 08:24:41.525393 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 20 08:24:41 crc kubenswrapper[4948]: I0220 08:24:41.525454 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 20 08:24:42 crc kubenswrapper[4948]: I0220 08:24:42.274283 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 20 08:24:42 crc kubenswrapper[4948]: I0220 08:24:42.304732 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 20 08:24:42 crc kubenswrapper[4948]: I0220 08:24:42.532694 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 20 08:24:42 crc kubenswrapper[4948]: I0220 08:24:42.533442 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 20 08:24:42 crc kubenswrapper[4948]: I0220 08:24:42.560309 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 08:24:42 crc kubenswrapper[4948]: I0220 08:24:42.560510 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 08:24:43 crc kubenswrapper[4948]: I0220 08:24:43.064452 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.453726 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.455931 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.456580 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.456749 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.464291 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.467104 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.533770 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.535720 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.540152 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.692155 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-t56xb"] Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.693547 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.739238 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-t56xb"] Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.788022 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-config\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.788127 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.788174 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.788235 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db22s\" (UniqueName: \"kubernetes.io/projected/c1847849-cdf4-4ea6-a540-c01503a003b2-kube-api-access-db22s\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.788269 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.788303 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.890027 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.890145 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db22s\" (UniqueName: \"kubernetes.io/projected/c1847849-cdf4-4ea6-a540-c01503a003b2-kube-api-access-db22s\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.890184 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.890225 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.890321 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-config\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.890369 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.890907 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.891376 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.891542 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-config\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.892040 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.892118 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:51 crc kubenswrapper[4948]: I0220 08:24:51.911195 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db22s\" (UniqueName: \"kubernetes.io/projected/c1847849-cdf4-4ea6-a540-c01503a003b2-kube-api-access-db22s\") pod \"dnsmasq-dns-89c5cd4d5-t56xb\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:52 crc kubenswrapper[4948]: I0220 08:24:52.023413 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:52 crc kubenswrapper[4948]: I0220 08:24:52.146725 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 20 08:24:52 crc kubenswrapper[4948]: I0220 08:24:52.570848 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-t56xb"] Feb 20 08:24:52 crc kubenswrapper[4948]: W0220 08:24:52.596817 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1847849_cdf4_4ea6_a540_c01503a003b2.slice/crio-4b70d018879f9b501eed3beab4a8c0781964989ca45e86dd4327f865ae78f6bf WatchSource:0}: Error finding container 4b70d018879f9b501eed3beab4a8c0781964989ca45e86dd4327f865ae78f6bf: Status 404 returned error can't find the container with id 4b70d018879f9b501eed3beab4a8c0781964989ca45e86dd4327f865ae78f6bf Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.148894 4948 generic.go:334] "Generic (PLEG): container finished" podID="c1847849-cdf4-4ea6-a540-c01503a003b2" containerID="6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a" exitCode=0 Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.149035 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" event={"ID":"c1847849-cdf4-4ea6-a540-c01503a003b2","Type":"ContainerDied","Data":"6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a"} Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.149527 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" event={"ID":"c1847849-cdf4-4ea6-a540-c01503a003b2","Type":"ContainerStarted","Data":"4b70d018879f9b501eed3beab4a8c0781964989ca45e86dd4327f865ae78f6bf"} Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.669343 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.669632 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="ceilometer-central-agent" containerID="cri-o://1630e56297ae1cd7ec343d1b51485df834ac4ef7ce2a756bf5a88dac19b42159" gracePeriod=30 Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.669746 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="ceilometer-notification-agent" containerID="cri-o://1bcad95c714fce375a1b0a71dbbd9424565d0d8570fd34397817043b9bfe2e88" gracePeriod=30 Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.669768 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="proxy-httpd" containerID="cri-o://430724be7dcee7dfd82861aa2d26928ec4bed184137610d057128b4eb3d23e14" gracePeriod=30 Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.669822 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="sg-core" containerID="cri-o://0574f1ab13d44d1f98f65e5fc2fb1727a44b35bb855f4f5336cf1da833d95b3a" gracePeriod=30 Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.701748 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Feb 20 08:24:53 crc kubenswrapper[4948]: I0220 08:24:53.971387 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.163603 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" event={"ID":"c1847849-cdf4-4ea6-a540-c01503a003b2","Type":"ContainerStarted","Data":"5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f"} Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.163933 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.166802 4948 generic.go:334] "Generic (PLEG): container finished" podID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerID="430724be7dcee7dfd82861aa2d26928ec4bed184137610d057128b4eb3d23e14" exitCode=0 Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.166825 4948 generic.go:334] "Generic (PLEG): container finished" podID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerID="0574f1ab13d44d1f98f65e5fc2fb1727a44b35bb855f4f5336cf1da833d95b3a" exitCode=2 Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.166833 4948 generic.go:334] "Generic (PLEG): container finished" podID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerID="1630e56297ae1cd7ec343d1b51485df834ac4ef7ce2a756bf5a88dac19b42159" exitCode=0 Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.166919 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerDied","Data":"430724be7dcee7dfd82861aa2d26928ec4bed184137610d057128b4eb3d23e14"} Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.167024 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerDied","Data":"0574f1ab13d44d1f98f65e5fc2fb1727a44b35bb855f4f5336cf1da833d95b3a"} Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.167046 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerDied","Data":"1630e56297ae1cd7ec343d1b51485df834ac4ef7ce2a756bf5a88dac19b42159"} Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.167289 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-log" containerID="cri-o://d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa" gracePeriod=30 Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.167328 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-api" containerID="cri-o://6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2" gracePeriod=30 Feb 20 08:24:54 crc kubenswrapper[4948]: I0220 08:24:54.196656 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" podStartSLOduration=3.196639074 podStartE2EDuration="3.196639074s" podCreationTimestamp="2026-02-20 08:24:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:54.194924773 +0000 UTC m=+1143.169419593" watchObservedRunningTime="2026-02-20 08:24:54.196639074 +0000 UTC m=+1143.171133894" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.034384 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.170178 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-combined-ca-bundle\") pod \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.170313 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-config-data\") pod \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.170360 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ll5q\" (UniqueName: \"kubernetes.io/projected/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-kube-api-access-2ll5q\") pod \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\" (UID: \"ad615fe7-b01a-4e4c-85b1-e028e5cdd866\") " Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.187328 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-kube-api-access-2ll5q" (OuterVolumeSpecName: "kube-api-access-2ll5q") pod "ad615fe7-b01a-4e4c-85b1-e028e5cdd866" (UID: "ad615fe7-b01a-4e4c-85b1-e028e5cdd866"). InnerVolumeSpecName "kube-api-access-2ll5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.190965 4948 generic.go:334] "Generic (PLEG): container finished" podID="4b950a35-d978-4338-88a6-ec96c9af048e" containerID="d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa" exitCode=143 Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.191136 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b950a35-d978-4338-88a6-ec96c9af048e","Type":"ContainerDied","Data":"d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa"} Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.192556 4948 generic.go:334] "Generic (PLEG): container finished" podID="ad615fe7-b01a-4e4c-85b1-e028e5cdd866" containerID="2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9" exitCode=137 Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.193653 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ad615fe7-b01a-4e4c-85b1-e028e5cdd866","Type":"ContainerDied","Data":"2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9"} Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.193703 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ad615fe7-b01a-4e4c-85b1-e028e5cdd866","Type":"ContainerDied","Data":"2e2fb5966e87b5f5075085440ad9aaba0868d3b137c6af57f293094a97faf543"} Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.193717 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.193722 4948 scope.go:117] "RemoveContainer" containerID="2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.232366 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad615fe7-b01a-4e4c-85b1-e028e5cdd866" (UID: "ad615fe7-b01a-4e4c-85b1-e028e5cdd866"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.252340 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-config-data" (OuterVolumeSpecName: "config-data") pod "ad615fe7-b01a-4e4c-85b1-e028e5cdd866" (UID: "ad615fe7-b01a-4e4c-85b1-e028e5cdd866"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.257484 4948 scope.go:117] "RemoveContainer" containerID="2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9" Feb 20 08:24:55 crc kubenswrapper[4948]: E0220 08:24:55.257830 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9\": container with ID starting with 2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9 not found: ID does not exist" containerID="2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.257860 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9"} err="failed to get container status \"2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9\": rpc error: code = NotFound desc = could not find container \"2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9\": container with ID starting with 2242b79213417225ca0b9ded36fc0195241950291c46926c9dff4400bf7ca4e9 not found: ID does not exist" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.274448 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.274514 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.274530 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ll5q\" (UniqueName: \"kubernetes.io/projected/ad615fe7-b01a-4e4c-85b1-e028e5cdd866-kube-api-access-2ll5q\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.324447 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.194:3000/\": dial tcp 10.217.0.194:3000: connect: connection refused" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.528702 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.536938 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.552131 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:55 crc kubenswrapper[4948]: E0220 08:24:55.553087 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad615fe7-b01a-4e4c-85b1-e028e5cdd866" containerName="nova-cell1-novncproxy-novncproxy" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.553134 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad615fe7-b01a-4e4c-85b1-e028e5cdd866" containerName="nova-cell1-novncproxy-novncproxy" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.553485 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad615fe7-b01a-4e4c-85b1-e028e5cdd866" containerName="nova-cell1-novncproxy-novncproxy" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.554151 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.558057 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.558185 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.558426 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.569476 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.682126 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.682195 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.682333 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.682426 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6tdc\" (UniqueName: \"kubernetes.io/projected/9c00f24b-fedb-4747-91c4-3d0b551a7288-kube-api-access-n6tdc\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.682463 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.734413 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad615fe7-b01a-4e4c-85b1-e028e5cdd866" path="/var/lib/kubelet/pods/ad615fe7-b01a-4e4c-85b1-e028e5cdd866/volumes" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.784639 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.784716 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.784856 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.784991 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6tdc\" (UniqueName: \"kubernetes.io/projected/9c00f24b-fedb-4747-91c4-3d0b551a7288-kube-api-access-n6tdc\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.785035 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.789503 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.790215 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.792409 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.792797 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c00f24b-fedb-4747-91c4-3d0b551a7288-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.804676 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6tdc\" (UniqueName: \"kubernetes.io/projected/9c00f24b-fedb-4747-91c4-3d0b551a7288-kube-api-access-n6tdc\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c00f24b-fedb-4747-91c4-3d0b551a7288\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:55 crc kubenswrapper[4948]: I0220 08:24:55.870682 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:24:56 crc kubenswrapper[4948]: I0220 08:24:56.344046 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 08:24:56 crc kubenswrapper[4948]: W0220 08:24:56.344830 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c00f24b_fedb_4747_91c4_3d0b551a7288.slice/crio-bffecdc61bc8fd8e92ede8f2c61eaf9d4b0abac0d945cb8a1261deb94cd475b9 WatchSource:0}: Error finding container bffecdc61bc8fd8e92ede8f2c61eaf9d4b0abac0d945cb8a1261deb94cd475b9: Status 404 returned error can't find the container with id bffecdc61bc8fd8e92ede8f2c61eaf9d4b0abac0d945cb8a1261deb94cd475b9 Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.238086 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9c00f24b-fedb-4747-91c4-3d0b551a7288","Type":"ContainerStarted","Data":"aed40c01186dd40a2c4f854e7c4f35be79dfb8528a952fc8f05c38778e541dff"} Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.238533 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9c00f24b-fedb-4747-91c4-3d0b551a7288","Type":"ContainerStarted","Data":"bffecdc61bc8fd8e92ede8f2c61eaf9d4b0abac0d945cb8a1261deb94cd475b9"} Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.242431 4948 generic.go:334] "Generic (PLEG): container finished" podID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerID="1bcad95c714fce375a1b0a71dbbd9424565d0d8570fd34397817043b9bfe2e88" exitCode=0 Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.242488 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerDied","Data":"1bcad95c714fce375a1b0a71dbbd9424565d0d8570fd34397817043b9bfe2e88"} Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.266100 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.266071646 podStartE2EDuration="2.266071646s" podCreationTimestamp="2026-02-20 08:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:24:57.255575583 +0000 UTC m=+1146.230070443" watchObservedRunningTime="2026-02-20 08:24:57.266071646 +0000 UTC m=+1146.240566476" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.521622 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.621895 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-config-data\") pod \"ee88c030-4657-480c-99a0-90ee4c2b247c\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.622008 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-combined-ca-bundle\") pod \"ee88c030-4657-480c-99a0-90ee4c2b247c\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.622077 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-run-httpd\") pod \"ee88c030-4657-480c-99a0-90ee4c2b247c\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.622101 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-log-httpd\") pod \"ee88c030-4657-480c-99a0-90ee4c2b247c\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.622157 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-scripts\") pod \"ee88c030-4657-480c-99a0-90ee4c2b247c\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.622206 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrzbr\" (UniqueName: \"kubernetes.io/projected/ee88c030-4657-480c-99a0-90ee4c2b247c-kube-api-access-rrzbr\") pod \"ee88c030-4657-480c-99a0-90ee4c2b247c\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.622311 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-sg-core-conf-yaml\") pod \"ee88c030-4657-480c-99a0-90ee4c2b247c\" (UID: \"ee88c030-4657-480c-99a0-90ee4c2b247c\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.623014 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ee88c030-4657-480c-99a0-90ee4c2b247c" (UID: "ee88c030-4657-480c-99a0-90ee4c2b247c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.626392 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ee88c030-4657-480c-99a0-90ee4c2b247c" (UID: "ee88c030-4657-480c-99a0-90ee4c2b247c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.648931 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee88c030-4657-480c-99a0-90ee4c2b247c-kube-api-access-rrzbr" (OuterVolumeSpecName: "kube-api-access-rrzbr") pod "ee88c030-4657-480c-99a0-90ee4c2b247c" (UID: "ee88c030-4657-480c-99a0-90ee4c2b247c"). InnerVolumeSpecName "kube-api-access-rrzbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.662620 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-scripts" (OuterVolumeSpecName: "scripts") pod "ee88c030-4657-480c-99a0-90ee4c2b247c" (UID: "ee88c030-4657-480c-99a0-90ee4c2b247c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.682058 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ee88c030-4657-480c-99a0-90ee4c2b247c" (UID: "ee88c030-4657-480c-99a0-90ee4c2b247c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.721159 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.724416 4948 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.724449 4948 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee88c030-4657-480c-99a0-90ee4c2b247c-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.724460 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.724471 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrzbr\" (UniqueName: \"kubernetes.io/projected/ee88c030-4657-480c-99a0-90ee4c2b247c-kube-api-access-rrzbr\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.724483 4948 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.757108 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee88c030-4657-480c-99a0-90ee4c2b247c" (UID: "ee88c030-4657-480c-99a0-90ee4c2b247c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.781184 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-config-data" (OuterVolumeSpecName: "config-data") pod "ee88c030-4657-480c-99a0-90ee4c2b247c" (UID: "ee88c030-4657-480c-99a0-90ee4c2b247c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.826684 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96ggh\" (UniqueName: \"kubernetes.io/projected/4b950a35-d978-4338-88a6-ec96c9af048e-kube-api-access-96ggh\") pod \"4b950a35-d978-4338-88a6-ec96c9af048e\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.826908 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-config-data\") pod \"4b950a35-d978-4338-88a6-ec96c9af048e\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.826938 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b950a35-d978-4338-88a6-ec96c9af048e-logs\") pod \"4b950a35-d978-4338-88a6-ec96c9af048e\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.827029 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-combined-ca-bundle\") pod \"4b950a35-d978-4338-88a6-ec96c9af048e\" (UID: \"4b950a35-d978-4338-88a6-ec96c9af048e\") " Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.827642 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.827667 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee88c030-4657-480c-99a0-90ee4c2b247c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.828520 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b950a35-d978-4338-88a6-ec96c9af048e-logs" (OuterVolumeSpecName: "logs") pod "4b950a35-d978-4338-88a6-ec96c9af048e" (UID: "4b950a35-d978-4338-88a6-ec96c9af048e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.834165 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b950a35-d978-4338-88a6-ec96c9af048e-kube-api-access-96ggh" (OuterVolumeSpecName: "kube-api-access-96ggh") pod "4b950a35-d978-4338-88a6-ec96c9af048e" (UID: "4b950a35-d978-4338-88a6-ec96c9af048e"). InnerVolumeSpecName "kube-api-access-96ggh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.862355 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-config-data" (OuterVolumeSpecName: "config-data") pod "4b950a35-d978-4338-88a6-ec96c9af048e" (UID: "4b950a35-d978-4338-88a6-ec96c9af048e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.865900 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b950a35-d978-4338-88a6-ec96c9af048e" (UID: "4b950a35-d978-4338-88a6-ec96c9af048e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.930086 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96ggh\" (UniqueName: \"kubernetes.io/projected/4b950a35-d978-4338-88a6-ec96c9af048e-kube-api-access-96ggh\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.930122 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.930135 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b950a35-d978-4338-88a6-ec96c9af048e-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:57 crc kubenswrapper[4948]: I0220 08:24:57.930145 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b950a35-d978-4338-88a6-ec96c9af048e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.263958 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.264007 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee88c030-4657-480c-99a0-90ee4c2b247c","Type":"ContainerDied","Data":"7a7f2ca4a85292e0cdfdd80686c701b75dbf3b7d4a7696caa3251ac73f926e4a"} Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.265013 4948 scope.go:117] "RemoveContainer" containerID="430724be7dcee7dfd82861aa2d26928ec4bed184137610d057128b4eb3d23e14" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.269247 4948 generic.go:334] "Generic (PLEG): container finished" podID="4b950a35-d978-4338-88a6-ec96c9af048e" containerID="6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2" exitCode=0 Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.269303 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.269362 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b950a35-d978-4338-88a6-ec96c9af048e","Type":"ContainerDied","Data":"6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2"} Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.269422 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b950a35-d978-4338-88a6-ec96c9af048e","Type":"ContainerDied","Data":"d66a33a2d7dbdb39f5f061898cb4a5149bc962e88f36c5fc69dae56ca6bdd17a"} Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.326952 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.336206 4948 scope.go:117] "RemoveContainer" containerID="0574f1ab13d44d1f98f65e5fc2fb1727a44b35bb855f4f5336cf1da833d95b3a" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.355733 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.365868 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.385245 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.396098 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:58 crc kubenswrapper[4948]: E0220 08:24:58.400138 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-log" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400181 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-log" Feb 20 08:24:58 crc kubenswrapper[4948]: E0220 08:24:58.400195 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="ceilometer-notification-agent" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400202 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="ceilometer-notification-agent" Feb 20 08:24:58 crc kubenswrapper[4948]: E0220 08:24:58.400212 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-api" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400218 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-api" Feb 20 08:24:58 crc kubenswrapper[4948]: E0220 08:24:58.400241 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="ceilometer-central-agent" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400246 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="ceilometer-central-agent" Feb 20 08:24:58 crc kubenswrapper[4948]: E0220 08:24:58.400258 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="proxy-httpd" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400264 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="proxy-httpd" Feb 20 08:24:58 crc kubenswrapper[4948]: E0220 08:24:58.400274 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="sg-core" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400281 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="sg-core" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400443 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-api" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400454 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="ceilometer-central-agent" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400470 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="sg-core" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400484 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="ceilometer-notification-agent" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400494 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" containerName="nova-api-log" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400506 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" containerName="proxy-httpd" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.400518 4948 scope.go:117] "RemoveContainer" containerID="1bcad95c714fce375a1b0a71dbbd9424565d0d8570fd34397817043b9bfe2e88" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.401424 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.403749 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.403871 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.403879 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.405955 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.414221 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.416693 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.426778 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.427127 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.427267 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.432509 4948 scope.go:117] "RemoveContainer" containerID="1630e56297ae1cd7ec343d1b51485df834ac4ef7ce2a756bf5a88dac19b42159" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.464283 4948 scope.go:117] "RemoveContainer" containerID="6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.483954 4948 scope.go:117] "RemoveContainer" containerID="d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.501096 4948 scope.go:117] "RemoveContainer" containerID="6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2" Feb 20 08:24:58 crc kubenswrapper[4948]: E0220 08:24:58.501457 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2\": container with ID starting with 6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2 not found: ID does not exist" containerID="6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.501504 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2"} err="failed to get container status \"6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2\": rpc error: code = NotFound desc = could not find container \"6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2\": container with ID starting with 6ef713a538c9adc323b40a21aaf7370bc2c92d050478ddd9bfe5e00490dfc4a2 not found: ID does not exist" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.501535 4948 scope.go:117] "RemoveContainer" containerID="d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa" Feb 20 08:24:58 crc kubenswrapper[4948]: E0220 08:24:58.501832 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa\": container with ID starting with d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa not found: ID does not exist" containerID="d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.501855 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa"} err="failed to get container status \"d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa\": rpc error: code = NotFound desc = could not find container \"d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa\": container with ID starting with d0e4271ae345ef60d4741e0350d3336cff2d587f6e410c83f608d1092728d8fa not found: ID does not exist" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.540912 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541019 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61657c77-bb29-4241-9d88-5e13bccd3f4e-logs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541056 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-config-data\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541082 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541343 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541418 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-scripts\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541570 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-public-tls-certs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541624 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541662 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-run-httpd\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541682 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwwjr\" (UniqueName: \"kubernetes.io/projected/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-kube-api-access-mwwjr\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541717 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr8fh\" (UniqueName: \"kubernetes.io/projected/61657c77-bb29-4241-9d88-5e13bccd3f4e-kube-api-access-zr8fh\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541735 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-config-data\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.541789 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-log-httpd\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.643807 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.643849 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-scripts\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.643904 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-public-tls-certs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.643929 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.643947 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwwjr\" (UniqueName: \"kubernetes.io/projected/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-kube-api-access-mwwjr\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.643965 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-run-httpd\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.643997 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr8fh\" (UniqueName: \"kubernetes.io/projected/61657c77-bb29-4241-9d88-5e13bccd3f4e-kube-api-access-zr8fh\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.644010 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-config-data\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.644034 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-log-httpd\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.644056 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.644500 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-run-httpd\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.644765 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61657c77-bb29-4241-9d88-5e13bccd3f4e-logs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.644797 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-config-data\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.644795 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-log-httpd\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.644815 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.645284 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61657c77-bb29-4241-9d88-5e13bccd3f4e-logs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.649386 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.649383 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-config-data\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.649590 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-public-tls-certs\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.650796 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.650984 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-config-data\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.654206 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-scripts\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.659767 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.665171 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.667124 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr8fh\" (UniqueName: \"kubernetes.io/projected/61657c77-bb29-4241-9d88-5e13bccd3f4e-kube-api-access-zr8fh\") pod \"nova-api-0\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.667650 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwwjr\" (UniqueName: \"kubernetes.io/projected/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-kube-api-access-mwwjr\") pod \"ceilometer-0\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " pod="openstack/ceilometer-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.718024 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:24:58 crc kubenswrapper[4948]: I0220 08:24:58.743047 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:24:59 crc kubenswrapper[4948]: I0220 08:24:59.222463 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:24:59 crc kubenswrapper[4948]: W0220 08:24:59.227679 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61657c77_bb29_4241_9d88_5e13bccd3f4e.slice/crio-cff172040120f907a742a7013b9d3e9198e3f80ccb81cd1545dba19de50f6adb WatchSource:0}: Error finding container cff172040120f907a742a7013b9d3e9198e3f80ccb81cd1545dba19de50f6adb: Status 404 returned error can't find the container with id cff172040120f907a742a7013b9d3e9198e3f80ccb81cd1545dba19de50f6adb Feb 20 08:24:59 crc kubenswrapper[4948]: W0220 08:24:59.231382 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe0e09bc_edd8_4ea9_8a59_ab630ed91cd8.slice/crio-48f0f09932c5f951969d8a7a57693456f1c18049262597223bdbf2f58c0804b2 WatchSource:0}: Error finding container 48f0f09932c5f951969d8a7a57693456f1c18049262597223bdbf2f58c0804b2: Status 404 returned error can't find the container with id 48f0f09932c5f951969d8a7a57693456f1c18049262597223bdbf2f58c0804b2 Feb 20 08:24:59 crc kubenswrapper[4948]: I0220 08:24:59.234325 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:24:59 crc kubenswrapper[4948]: I0220 08:24:59.293026 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerStarted","Data":"48f0f09932c5f951969d8a7a57693456f1c18049262597223bdbf2f58c0804b2"} Feb 20 08:24:59 crc kubenswrapper[4948]: I0220 08:24:59.296842 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61657c77-bb29-4241-9d88-5e13bccd3f4e","Type":"ContainerStarted","Data":"cff172040120f907a742a7013b9d3e9198e3f80ccb81cd1545dba19de50f6adb"} Feb 20 08:24:59 crc kubenswrapper[4948]: I0220 08:24:59.739604 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b950a35-d978-4338-88a6-ec96c9af048e" path="/var/lib/kubelet/pods/4b950a35-d978-4338-88a6-ec96c9af048e/volumes" Feb 20 08:24:59 crc kubenswrapper[4948]: I0220 08:24:59.741074 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee88c030-4657-480c-99a0-90ee4c2b247c" path="/var/lib/kubelet/pods/ee88c030-4657-480c-99a0-90ee4c2b247c/volumes" Feb 20 08:25:00 crc kubenswrapper[4948]: I0220 08:25:00.312275 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerStarted","Data":"d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a"} Feb 20 08:25:00 crc kubenswrapper[4948]: I0220 08:25:00.314794 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61657c77-bb29-4241-9d88-5e13bccd3f4e","Type":"ContainerStarted","Data":"185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc"} Feb 20 08:25:00 crc kubenswrapper[4948]: I0220 08:25:00.314842 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61657c77-bb29-4241-9d88-5e13bccd3f4e","Type":"ContainerStarted","Data":"38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af"} Feb 20 08:25:00 crc kubenswrapper[4948]: I0220 08:25:00.351775 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.351760001 podStartE2EDuration="2.351760001s" podCreationTimestamp="2026-02-20 08:24:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:25:00.340812587 +0000 UTC m=+1149.315307417" watchObservedRunningTime="2026-02-20 08:25:00.351760001 +0000 UTC m=+1149.326254821" Feb 20 08:25:00 crc kubenswrapper[4948]: I0220 08:25:00.871668 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:25:01 crc kubenswrapper[4948]: I0220 08:25:01.324918 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerStarted","Data":"e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0"} Feb 20 08:25:01 crc kubenswrapper[4948]: I0220 08:25:01.325328 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerStarted","Data":"43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542"} Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.026185 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.104027 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x22px"] Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.104282 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-x22px" podUID="538c1b6f-424d-48e3-a92b-7534fbf865fc" containerName="dnsmasq-dns" containerID="cri-o://3a9357bad2b3dca80dc83315104b35d29a1b980c7594ca0611bf4b94b36eb262" gracePeriod=10 Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.338038 4948 generic.go:334] "Generic (PLEG): container finished" podID="538c1b6f-424d-48e3-a92b-7534fbf865fc" containerID="3a9357bad2b3dca80dc83315104b35d29a1b980c7594ca0611bf4b94b36eb262" exitCode=0 Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.338091 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-x22px" event={"ID":"538c1b6f-424d-48e3-a92b-7534fbf865fc","Type":"ContainerDied","Data":"3a9357bad2b3dca80dc83315104b35d29a1b980c7594ca0611bf4b94b36eb262"} Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.637499 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.725439 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb\") pod \"538c1b6f-424d-48e3-a92b-7534fbf865fc\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.725588 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hh28w\" (UniqueName: \"kubernetes.io/projected/538c1b6f-424d-48e3-a92b-7534fbf865fc-kube-api-access-hh28w\") pod \"538c1b6f-424d-48e3-a92b-7534fbf865fc\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.725683 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-config\") pod \"538c1b6f-424d-48e3-a92b-7534fbf865fc\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.725720 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-swift-storage-0\") pod \"538c1b6f-424d-48e3-a92b-7534fbf865fc\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.725783 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-sb\") pod \"538c1b6f-424d-48e3-a92b-7534fbf865fc\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.725807 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-svc\") pod \"538c1b6f-424d-48e3-a92b-7534fbf865fc\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.734221 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/538c1b6f-424d-48e3-a92b-7534fbf865fc-kube-api-access-hh28w" (OuterVolumeSpecName: "kube-api-access-hh28w") pod "538c1b6f-424d-48e3-a92b-7534fbf865fc" (UID: "538c1b6f-424d-48e3-a92b-7534fbf865fc"). InnerVolumeSpecName "kube-api-access-hh28w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.778035 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "538c1b6f-424d-48e3-a92b-7534fbf865fc" (UID: "538c1b6f-424d-48e3-a92b-7534fbf865fc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.780815 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "538c1b6f-424d-48e3-a92b-7534fbf865fc" (UID: "538c1b6f-424d-48e3-a92b-7534fbf865fc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.784027 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "538c1b6f-424d-48e3-a92b-7534fbf865fc" (UID: "538c1b6f-424d-48e3-a92b-7534fbf865fc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:25:02 crc kubenswrapper[4948]: E0220 08:25:02.799387 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb podName:538c1b6f-424d-48e3-a92b-7534fbf865fc nodeName:}" failed. No retries permitted until 2026-02-20 08:25:03.299359908 +0000 UTC m=+1152.273854728 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ovsdbserver-nb" (UniqueName: "kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb") pod "538c1b6f-424d-48e3-a92b-7534fbf865fc" (UID: "538c1b6f-424d-48e3-a92b-7534fbf865fc") : error deleting /var/lib/kubelet/pods/538c1b6f-424d-48e3-a92b-7534fbf865fc/volume-subpaths: remove /var/lib/kubelet/pods/538c1b6f-424d-48e3-a92b-7534fbf865fc/volume-subpaths: no such file or directory Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.799670 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-config" (OuterVolumeSpecName: "config") pod "538c1b6f-424d-48e3-a92b-7534fbf865fc" (UID: "538c1b6f-424d-48e3-a92b-7534fbf865fc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.832857 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hh28w\" (UniqueName: \"kubernetes.io/projected/538c1b6f-424d-48e3-a92b-7534fbf865fc-kube-api-access-hh28w\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.832915 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.832929 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.832943 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:02 crc kubenswrapper[4948]: I0220 08:25:02.832986 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.342878 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb\") pod \"538c1b6f-424d-48e3-a92b-7534fbf865fc\" (UID: \"538c1b6f-424d-48e3-a92b-7534fbf865fc\") " Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.345787 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "538c1b6f-424d-48e3-a92b-7534fbf865fc" (UID: "538c1b6f-424d-48e3-a92b-7534fbf865fc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.358181 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-x22px" event={"ID":"538c1b6f-424d-48e3-a92b-7534fbf865fc","Type":"ContainerDied","Data":"ca84f60c8f59ce30814af7db2ac25fea0b071bd0c109ef4bdad869f451268b2b"} Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.358196 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-x22px" Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.358860 4948 scope.go:117] "RemoveContainer" containerID="3a9357bad2b3dca80dc83315104b35d29a1b980c7594ca0611bf4b94b36eb262" Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.362536 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerStarted","Data":"a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9"} Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.363379 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.387716 4948 scope.go:117] "RemoveContainer" containerID="70074ef7c3184ccb90626f2f932107dda3ac2287a8b654203232efa428fd6294" Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.421540 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.942159117 podStartE2EDuration="5.421504641s" podCreationTimestamp="2026-02-20 08:24:58 +0000 UTC" firstStartedPulling="2026-02-20 08:24:59.233563857 +0000 UTC m=+1148.208058677" lastFinishedPulling="2026-02-20 08:25:02.712909381 +0000 UTC m=+1151.687404201" observedRunningTime="2026-02-20 08:25:03.399280294 +0000 UTC m=+1152.373775114" watchObservedRunningTime="2026-02-20 08:25:03.421504641 +0000 UTC m=+1152.395999461" Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.429362 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x22px"] Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.440134 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-x22px"] Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.447058 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/538c1b6f-424d-48e3-a92b-7534fbf865fc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:03 crc kubenswrapper[4948]: I0220 08:25:03.733611 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="538c1b6f-424d-48e3-a92b-7534fbf865fc" path="/var/lib/kubelet/pods/538c1b6f-424d-48e3-a92b-7534fbf865fc/volumes" Feb 20 08:25:05 crc kubenswrapper[4948]: I0220 08:25:05.871490 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:25:05 crc kubenswrapper[4948]: I0220 08:25:05.893947 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.410893 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.608958 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-mg24f"] Feb 20 08:25:06 crc kubenswrapper[4948]: E0220 08:25:06.609462 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="538c1b6f-424d-48e3-a92b-7534fbf865fc" containerName="dnsmasq-dns" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.609477 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="538c1b6f-424d-48e3-a92b-7534fbf865fc" containerName="dnsmasq-dns" Feb 20 08:25:06 crc kubenswrapper[4948]: E0220 08:25:06.609499 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="538c1b6f-424d-48e3-a92b-7534fbf865fc" containerName="init" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.609507 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="538c1b6f-424d-48e3-a92b-7534fbf865fc" containerName="init" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.609737 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="538c1b6f-424d-48e3-a92b-7534fbf865fc" containerName="dnsmasq-dns" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.610514 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.614088 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.614563 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.623939 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mg24f"] Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.715182 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf5d4\" (UniqueName: \"kubernetes.io/projected/cf549f46-7597-4399-addc-3bf3f5b63f99-kube-api-access-mf5d4\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.715253 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-scripts\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.715277 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-config-data\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.715358 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.816683 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-scripts\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.816745 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-config-data\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.816845 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.817037 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf5d4\" (UniqueName: \"kubernetes.io/projected/cf549f46-7597-4399-addc-3bf3f5b63f99-kube-api-access-mf5d4\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.823056 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-scripts\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.830672 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-config-data\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.831063 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.840624 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf5d4\" (UniqueName: \"kubernetes.io/projected/cf549f46-7597-4399-addc-3bf3f5b63f99-kube-api-access-mf5d4\") pod \"nova-cell1-cell-mapping-mg24f\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:06 crc kubenswrapper[4948]: I0220 08:25:06.950873 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:07 crc kubenswrapper[4948]: I0220 08:25:07.338229 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mg24f"] Feb 20 08:25:07 crc kubenswrapper[4948]: I0220 08:25:07.403709 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mg24f" event={"ID":"cf549f46-7597-4399-addc-3bf3f5b63f99","Type":"ContainerStarted","Data":"f221fd286de80b0b0916709f7cbe440e6ef10e15899275765ef094ff94516225"} Feb 20 08:25:08 crc kubenswrapper[4948]: I0220 08:25:08.416705 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mg24f" event={"ID":"cf549f46-7597-4399-addc-3bf3f5b63f99","Type":"ContainerStarted","Data":"d68a0fb50da62e47bddf140b8e1628a7abf7e9ac68f7c4c4efaa6f9ccf62f179"} Feb 20 08:25:08 crc kubenswrapper[4948]: I0220 08:25:08.436714 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-mg24f" podStartSLOduration=2.436694909 podStartE2EDuration="2.436694909s" podCreationTimestamp="2026-02-20 08:25:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:25:08.434721291 +0000 UTC m=+1157.409216111" watchObservedRunningTime="2026-02-20 08:25:08.436694909 +0000 UTC m=+1157.411189729" Feb 20 08:25:08 crc kubenswrapper[4948]: I0220 08:25:08.718808 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 08:25:08 crc kubenswrapper[4948]: I0220 08:25:08.718849 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 08:25:09 crc kubenswrapper[4948]: I0220 08:25:09.739886 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 20 08:25:09 crc kubenswrapper[4948]: I0220 08:25:09.740676 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 08:25:12 crc kubenswrapper[4948]: I0220 08:25:12.452094 4948 generic.go:334] "Generic (PLEG): container finished" podID="cf549f46-7597-4399-addc-3bf3f5b63f99" containerID="d68a0fb50da62e47bddf140b8e1628a7abf7e9ac68f7c4c4efaa6f9ccf62f179" exitCode=0 Feb 20 08:25:12 crc kubenswrapper[4948]: I0220 08:25:12.452189 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mg24f" event={"ID":"cf549f46-7597-4399-addc-3bf3f5b63f99","Type":"ContainerDied","Data":"d68a0fb50da62e47bddf140b8e1628a7abf7e9ac68f7c4c4efaa6f9ccf62f179"} Feb 20 08:25:13 crc kubenswrapper[4948]: I0220 08:25:13.884118 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:13 crc kubenswrapper[4948]: I0220 08:25:13.974964 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf5d4\" (UniqueName: \"kubernetes.io/projected/cf549f46-7597-4399-addc-3bf3f5b63f99-kube-api-access-mf5d4\") pod \"cf549f46-7597-4399-addc-3bf3f5b63f99\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " Feb 20 08:25:13 crc kubenswrapper[4948]: I0220 08:25:13.975066 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-config-data\") pod \"cf549f46-7597-4399-addc-3bf3f5b63f99\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " Feb 20 08:25:13 crc kubenswrapper[4948]: I0220 08:25:13.975117 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-combined-ca-bundle\") pod \"cf549f46-7597-4399-addc-3bf3f5b63f99\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " Feb 20 08:25:13 crc kubenswrapper[4948]: I0220 08:25:13.975195 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-scripts\") pod \"cf549f46-7597-4399-addc-3bf3f5b63f99\" (UID: \"cf549f46-7597-4399-addc-3bf3f5b63f99\") " Feb 20 08:25:13 crc kubenswrapper[4948]: I0220 08:25:13.982448 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf549f46-7597-4399-addc-3bf3f5b63f99-kube-api-access-mf5d4" (OuterVolumeSpecName: "kube-api-access-mf5d4") pod "cf549f46-7597-4399-addc-3bf3f5b63f99" (UID: "cf549f46-7597-4399-addc-3bf3f5b63f99"). InnerVolumeSpecName "kube-api-access-mf5d4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:25:13 crc kubenswrapper[4948]: I0220 08:25:13.987695 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-scripts" (OuterVolumeSpecName: "scripts") pod "cf549f46-7597-4399-addc-3bf3f5b63f99" (UID: "cf549f46-7597-4399-addc-3bf3f5b63f99"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.002484 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-config-data" (OuterVolumeSpecName: "config-data") pod "cf549f46-7597-4399-addc-3bf3f5b63f99" (UID: "cf549f46-7597-4399-addc-3bf3f5b63f99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.005161 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf549f46-7597-4399-addc-3bf3f5b63f99" (UID: "cf549f46-7597-4399-addc-3bf3f5b63f99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.077535 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf5d4\" (UniqueName: \"kubernetes.io/projected/cf549f46-7597-4399-addc-3bf3f5b63f99-kube-api-access-mf5d4\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.077566 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.077576 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.077584 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf549f46-7597-4399-addc-3bf3f5b63f99-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.470859 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mg24f" event={"ID":"cf549f46-7597-4399-addc-3bf3f5b63f99","Type":"ContainerDied","Data":"f221fd286de80b0b0916709f7cbe440e6ef10e15899275765ef094ff94516225"} Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.470893 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f221fd286de80b0b0916709f7cbe440e6ef10e15899275765ef094ff94516225" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.470960 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mg24f" Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.675110 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.675735 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a76350cb-da45-403e-8934-ef53b257ea27" containerName="nova-scheduler-scheduler" containerID="cri-o://7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787" gracePeriod=30 Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.695550 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.695829 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-log" containerID="cri-o://38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af" gracePeriod=30 Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.695999 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-api" containerID="cri-o://185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc" gracePeriod=30 Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.716701 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.717185 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-log" containerID="cri-o://63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf" gracePeriod=30 Feb 20 08:25:14 crc kubenswrapper[4948]: I0220 08:25:14.717302 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-metadata" containerID="cri-o://22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9" gracePeriod=30 Feb 20 08:25:15 crc kubenswrapper[4948]: I0220 08:25:15.479841 4948 generic.go:334] "Generic (PLEG): container finished" podID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerID="38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af" exitCode=143 Feb 20 08:25:15 crc kubenswrapper[4948]: I0220 08:25:15.479917 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61657c77-bb29-4241-9d88-5e13bccd3f4e","Type":"ContainerDied","Data":"38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af"} Feb 20 08:25:15 crc kubenswrapper[4948]: I0220 08:25:15.482020 4948 generic.go:334] "Generic (PLEG): container finished" podID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerID="63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf" exitCode=143 Feb 20 08:25:15 crc kubenswrapper[4948]: I0220 08:25:15.482062 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"659ea49b-7e8f-4c95-ac69-ec743d0f95ed","Type":"ContainerDied","Data":"63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf"} Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.097429 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.217952 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-combined-ca-bundle\") pod \"a76350cb-da45-403e-8934-ef53b257ea27\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.218064 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-config-data\") pod \"a76350cb-da45-403e-8934-ef53b257ea27\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.218179 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb4nn\" (UniqueName: \"kubernetes.io/projected/a76350cb-da45-403e-8934-ef53b257ea27-kube-api-access-cb4nn\") pod \"a76350cb-da45-403e-8934-ef53b257ea27\" (UID: \"a76350cb-da45-403e-8934-ef53b257ea27\") " Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.236297 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a76350cb-da45-403e-8934-ef53b257ea27-kube-api-access-cb4nn" (OuterVolumeSpecName: "kube-api-access-cb4nn") pod "a76350cb-da45-403e-8934-ef53b257ea27" (UID: "a76350cb-da45-403e-8934-ef53b257ea27"). InnerVolumeSpecName "kube-api-access-cb4nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.256980 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-config-data" (OuterVolumeSpecName: "config-data") pod "a76350cb-da45-403e-8934-ef53b257ea27" (UID: "a76350cb-da45-403e-8934-ef53b257ea27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.281489 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a76350cb-da45-403e-8934-ef53b257ea27" (UID: "a76350cb-da45-403e-8934-ef53b257ea27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.320138 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.320299 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76350cb-da45-403e-8934-ef53b257ea27-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.320358 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb4nn\" (UniqueName: \"kubernetes.io/projected/a76350cb-da45-403e-8934-ef53b257ea27-kube-api-access-cb4nn\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.492265 4948 generic.go:334] "Generic (PLEG): container finished" podID="a76350cb-da45-403e-8934-ef53b257ea27" containerID="7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787" exitCode=0 Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.492308 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.492327 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a76350cb-da45-403e-8934-ef53b257ea27","Type":"ContainerDied","Data":"7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787"} Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.493537 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a76350cb-da45-403e-8934-ef53b257ea27","Type":"ContainerDied","Data":"6fbbcc32e0efcd622515f0c849bc2e7cae40139cc47cc3aeb9187b65bfe28ca5"} Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.493575 4948 scope.go:117] "RemoveContainer" containerID="7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.527126 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.530418 4948 scope.go:117] "RemoveContainer" containerID="7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787" Feb 20 08:25:16 crc kubenswrapper[4948]: E0220 08:25:16.530831 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787\": container with ID starting with 7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787 not found: ID does not exist" containerID="7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.530867 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787"} err="failed to get container status \"7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787\": rpc error: code = NotFound desc = could not find container \"7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787\": container with ID starting with 7fdf44e32062d9617c830886caa9e40ee3055d0b238b4f4ec53507c904d02787 not found: ID does not exist" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.537292 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.561359 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:25:16 crc kubenswrapper[4948]: E0220 08:25:16.561823 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a76350cb-da45-403e-8934-ef53b257ea27" containerName="nova-scheduler-scheduler" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.561842 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="a76350cb-da45-403e-8934-ef53b257ea27" containerName="nova-scheduler-scheduler" Feb 20 08:25:16 crc kubenswrapper[4948]: E0220 08:25:16.561867 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf549f46-7597-4399-addc-3bf3f5b63f99" containerName="nova-manage" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.561875 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf549f46-7597-4399-addc-3bf3f5b63f99" containerName="nova-manage" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.562174 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="a76350cb-da45-403e-8934-ef53b257ea27" containerName="nova-scheduler-scheduler" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.562226 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf549f46-7597-4399-addc-3bf3f5b63f99" containerName="nova-manage" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.563165 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.565324 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.582577 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.625241 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js86x\" (UniqueName: \"kubernetes.io/projected/f114bc41-01a5-4955-97eb-7fcf139cc5a9-kube-api-access-js86x\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.625280 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f114bc41-01a5-4955-97eb-7fcf139cc5a9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.625351 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f114bc41-01a5-4955-97eb-7fcf139cc5a9-config-data\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.726551 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f114bc41-01a5-4955-97eb-7fcf139cc5a9-config-data\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.726702 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js86x\" (UniqueName: \"kubernetes.io/projected/f114bc41-01a5-4955-97eb-7fcf139cc5a9-kube-api-access-js86x\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.726732 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f114bc41-01a5-4955-97eb-7fcf139cc5a9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.732713 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f114bc41-01a5-4955-97eb-7fcf139cc5a9-config-data\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.732730 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f114bc41-01a5-4955-97eb-7fcf139cc5a9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.754832 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js86x\" (UniqueName: \"kubernetes.io/projected/f114bc41-01a5-4955-97eb-7fcf139cc5a9-kube-api-access-js86x\") pod \"nova-scheduler-0\" (UID: \"f114bc41-01a5-4955-97eb-7fcf139cc5a9\") " pod="openstack/nova-scheduler-0" Feb 20 08:25:16 crc kubenswrapper[4948]: I0220 08:25:16.880688 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 08:25:17 crc kubenswrapper[4948]: I0220 08:25:17.324034 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 08:25:17 crc kubenswrapper[4948]: W0220 08:25:17.329498 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf114bc41_01a5_4955_97eb_7fcf139cc5a9.slice/crio-7221e125e11e0334cd672a971302d9b4b074d8307c59383d81cf365b5c4aac21 WatchSource:0}: Error finding container 7221e125e11e0334cd672a971302d9b4b074d8307c59383d81cf365b5c4aac21: Status 404 returned error can't find the container with id 7221e125e11e0334cd672a971302d9b4b074d8307c59383d81cf365b5c4aac21 Feb 20 08:25:17 crc kubenswrapper[4948]: I0220 08:25:17.507141 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f114bc41-01a5-4955-97eb-7fcf139cc5a9","Type":"ContainerStarted","Data":"7221e125e11e0334cd672a971302d9b4b074d8307c59383d81cf365b5c4aac21"} Feb 20 08:25:17 crc kubenswrapper[4948]: I0220 08:25:17.741700 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a76350cb-da45-403e-8934-ef53b257ea27" path="/var/lib/kubelet/pods/a76350cb-da45-403e-8934-ef53b257ea27/volumes" Feb 20 08:25:17 crc kubenswrapper[4948]: I0220 08:25:17.871244 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": read tcp 10.217.0.2:45064->10.217.0.198:8775: read: connection reset by peer" Feb 20 08:25:17 crc kubenswrapper[4948]: I0220 08:25:17.871257 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": read tcp 10.217.0.2:45078->10.217.0.198:8775: read: connection reset by peer" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.417927 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.423986 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.463320 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-config-data\") pod \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.463374 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-combined-ca-bundle\") pod \"61657c77-bb29-4241-9d88-5e13bccd3f4e\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464242 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr8fh\" (UniqueName: \"kubernetes.io/projected/61657c77-bb29-4241-9d88-5e13bccd3f4e-kube-api-access-zr8fh\") pod \"61657c77-bb29-4241-9d88-5e13bccd3f4e\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464288 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-config-data\") pod \"61657c77-bb29-4241-9d88-5e13bccd3f4e\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464355 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-logs\") pod \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464398 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-internal-tls-certs\") pod \"61657c77-bb29-4241-9d88-5e13bccd3f4e\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464440 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-combined-ca-bundle\") pod \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464469 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-public-tls-certs\") pod \"61657c77-bb29-4241-9d88-5e13bccd3f4e\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464523 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7ch5\" (UniqueName: \"kubernetes.io/projected/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-kube-api-access-p7ch5\") pod \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464553 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61657c77-bb29-4241-9d88-5e13bccd3f4e-logs\") pod \"61657c77-bb29-4241-9d88-5e13bccd3f4e\" (UID: \"61657c77-bb29-4241-9d88-5e13bccd3f4e\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.464592 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-nova-metadata-tls-certs\") pod \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\" (UID: \"659ea49b-7e8f-4c95-ac69-ec743d0f95ed\") " Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.470011 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61657c77-bb29-4241-9d88-5e13bccd3f4e-kube-api-access-zr8fh" (OuterVolumeSpecName: "kube-api-access-zr8fh") pod "61657c77-bb29-4241-9d88-5e13bccd3f4e" (UID: "61657c77-bb29-4241-9d88-5e13bccd3f4e"). InnerVolumeSpecName "kube-api-access-zr8fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.470546 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61657c77-bb29-4241-9d88-5e13bccd3f4e-logs" (OuterVolumeSpecName: "logs") pod "61657c77-bb29-4241-9d88-5e13bccd3f4e" (UID: "61657c77-bb29-4241-9d88-5e13bccd3f4e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.472119 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-logs" (OuterVolumeSpecName: "logs") pod "659ea49b-7e8f-4c95-ac69-ec743d0f95ed" (UID: "659ea49b-7e8f-4c95-ac69-ec743d0f95ed"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.491826 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-kube-api-access-p7ch5" (OuterVolumeSpecName: "kube-api-access-p7ch5") pod "659ea49b-7e8f-4c95-ac69-ec743d0f95ed" (UID: "659ea49b-7e8f-4c95-ac69-ec743d0f95ed"). InnerVolumeSpecName "kube-api-access-p7ch5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.529769 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-config-data" (OuterVolumeSpecName: "config-data") pod "659ea49b-7e8f-4c95-ac69-ec743d0f95ed" (UID: "659ea49b-7e8f-4c95-ac69-ec743d0f95ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.530054 4948 generic.go:334] "Generic (PLEG): container finished" podID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerID="22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9" exitCode=0 Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.530120 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.530196 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"659ea49b-7e8f-4c95-ac69-ec743d0f95ed","Type":"ContainerDied","Data":"22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9"} Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.530229 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"659ea49b-7e8f-4c95-ac69-ec743d0f95ed","Type":"ContainerDied","Data":"aaa4c4ee0564a915596564697d18943937054d1b42224076a88c2793a4afd7d0"} Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.530250 4948 scope.go:117] "RemoveContainer" containerID="22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.533752 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f114bc41-01a5-4955-97eb-7fcf139cc5a9","Type":"ContainerStarted","Data":"c760bba370b0c11b1caf2e537a3240838f2824510f92c38d77c5ed335aec7b26"} Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.542246 4948 generic.go:334] "Generic (PLEG): container finished" podID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerID="185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc" exitCode=0 Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.542316 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61657c77-bb29-4241-9d88-5e13bccd3f4e","Type":"ContainerDied","Data":"185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc"} Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.542346 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61657c77-bb29-4241-9d88-5e13bccd3f4e","Type":"ContainerDied","Data":"cff172040120f907a742a7013b9d3e9198e3f80ccb81cd1545dba19de50f6adb"} Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.542408 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.548196 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61657c77-bb29-4241-9d88-5e13bccd3f4e" (UID: "61657c77-bb29-4241-9d88-5e13bccd3f4e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.558590 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.558568762 podStartE2EDuration="2.558568762s" podCreationTimestamp="2026-02-20 08:25:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:25:18.554158756 +0000 UTC m=+1167.528653596" watchObservedRunningTime="2026-02-20 08:25:18.558568762 +0000 UTC m=+1167.533063582" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.558888 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "659ea49b-7e8f-4c95-ac69-ec743d0f95ed" (UID: "659ea49b-7e8f-4c95-ac69-ec743d0f95ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.566829 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.566860 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zr8fh\" (UniqueName: \"kubernetes.io/projected/61657c77-bb29-4241-9d88-5e13bccd3f4e-kube-api-access-zr8fh\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.566875 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.566887 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.566900 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7ch5\" (UniqueName: \"kubernetes.io/projected/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-kube-api-access-p7ch5\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.566911 4948 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61657c77-bb29-4241-9d88-5e13bccd3f4e-logs\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.566921 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.596542 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-config-data" (OuterVolumeSpecName: "config-data") pod "61657c77-bb29-4241-9d88-5e13bccd3f4e" (UID: "61657c77-bb29-4241-9d88-5e13bccd3f4e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.597665 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "61657c77-bb29-4241-9d88-5e13bccd3f4e" (UID: "61657c77-bb29-4241-9d88-5e13bccd3f4e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.599545 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "61657c77-bb29-4241-9d88-5e13bccd3f4e" (UID: "61657c77-bb29-4241-9d88-5e13bccd3f4e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.623883 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "659ea49b-7e8f-4c95-ac69-ec743d0f95ed" (UID: "659ea49b-7e8f-4c95-ac69-ec743d0f95ed"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.626180 4948 scope.go:117] "RemoveContainer" containerID="63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.645144 4948 scope.go:117] "RemoveContainer" containerID="22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9" Feb 20 08:25:18 crc kubenswrapper[4948]: E0220 08:25:18.646478 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9\": container with ID starting with 22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9 not found: ID does not exist" containerID="22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.646586 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9"} err="failed to get container status \"22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9\": rpc error: code = NotFound desc = could not find container \"22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9\": container with ID starting with 22196e9ec797dbe697aeeef9301c64792cbe1db93384c86a9e5389ca3fe652c9 not found: ID does not exist" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.646718 4948 scope.go:117] "RemoveContainer" containerID="63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf" Feb 20 08:25:18 crc kubenswrapper[4948]: E0220 08:25:18.647183 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf\": container with ID starting with 63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf not found: ID does not exist" containerID="63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.647258 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf"} err="failed to get container status \"63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf\": rpc error: code = NotFound desc = could not find container \"63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf\": container with ID starting with 63251b467317ee4b7966810a99bfea1f41b91c2ea1703852741d551304c8bdaf not found: ID does not exist" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.647371 4948 scope.go:117] "RemoveContainer" containerID="185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.663963 4948 scope.go:117] "RemoveContainer" containerID="38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.668851 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.668879 4948 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.668889 4948 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/61657c77-bb29-4241-9d88-5e13bccd3f4e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.668899 4948 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/659ea49b-7e8f-4c95-ac69-ec743d0f95ed-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.685050 4948 scope.go:117] "RemoveContainer" containerID="185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc" Feb 20 08:25:18 crc kubenswrapper[4948]: E0220 08:25:18.685876 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc\": container with ID starting with 185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc not found: ID does not exist" containerID="185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.685942 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc"} err="failed to get container status \"185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc\": rpc error: code = NotFound desc = could not find container \"185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc\": container with ID starting with 185d2ffc7312a2c6e5587955ab1e83610f6970f92e1e567908cad389309483bc not found: ID does not exist" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.686006 4948 scope.go:117] "RemoveContainer" containerID="38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af" Feb 20 08:25:18 crc kubenswrapper[4948]: E0220 08:25:18.686358 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af\": container with ID starting with 38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af not found: ID does not exist" containerID="38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.686401 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af"} err="failed to get container status \"38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af\": rpc error: code = NotFound desc = could not find container \"38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af\": container with ID starting with 38deb2b3c4145711eaa3d46abb6b2827da68608a05ee0a36a26658e34f7dc7af not found: ID does not exist" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.880234 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.894444 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.905916 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:25:18 crc kubenswrapper[4948]: E0220 08:25:18.906420 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-metadata" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.906447 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-metadata" Feb 20 08:25:18 crc kubenswrapper[4948]: E0220 08:25:18.906470 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-log" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.906480 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-log" Feb 20 08:25:18 crc kubenswrapper[4948]: E0220 08:25:18.906493 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-api" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.906501 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-api" Feb 20 08:25:18 crc kubenswrapper[4948]: E0220 08:25:18.906531 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-log" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.906540 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-log" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.906733 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-log" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.906761 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" containerName="nova-metadata-metadata" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.906775 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-log" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.906791 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" containerName="nova-api-api" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.908046 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.911192 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.912551 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.915696 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.924333 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.936099 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.944383 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.946426 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.949254 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.949720 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.960968 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.972277 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974487 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-public-tls-certs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974535 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974585 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-logs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974640 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974679 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59965363-9971-4130-bb5a-5fcf5a44e2d5-logs\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974699 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974738 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhr8w\" (UniqueName: \"kubernetes.io/projected/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-kube-api-access-zhr8w\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974806 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974848 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-config-data\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974885 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-config-data\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:18 crc kubenswrapper[4948]: I0220 08:25:18.974916 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sqtf\" (UniqueName: \"kubernetes.io/projected/59965363-9971-4130-bb5a-5fcf5a44e2d5-kube-api-access-5sqtf\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.076982 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-public-tls-certs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077041 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077093 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-logs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077130 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077167 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59965363-9971-4130-bb5a-5fcf5a44e2d5-logs\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077189 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077229 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhr8w\" (UniqueName: \"kubernetes.io/projected/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-kube-api-access-zhr8w\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077293 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077327 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-config-data\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077363 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-config-data\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077396 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sqtf\" (UniqueName: \"kubernetes.io/projected/59965363-9971-4130-bb5a-5fcf5a44e2d5-kube-api-access-5sqtf\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.077543 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-logs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.078098 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59965363-9971-4130-bb5a-5fcf5a44e2d5-logs\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.080016 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-public-tls-certs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.080837 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-config-data\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.080909 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.081584 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-config-data\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.081910 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.083376 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.084056 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/59965363-9971-4130-bb5a-5fcf5a44e2d5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.093019 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhr8w\" (UniqueName: \"kubernetes.io/projected/354f9b2f-9eaa-4e47-b214-0fd895f90c4c-kube-api-access-zhr8w\") pod \"nova-api-0\" (UID: \"354f9b2f-9eaa-4e47-b214-0fd895f90c4c\") " pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.094819 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sqtf\" (UniqueName: \"kubernetes.io/projected/59965363-9971-4130-bb5a-5fcf5a44e2d5-kube-api-access-5sqtf\") pod \"nova-metadata-0\" (UID: \"59965363-9971-4130-bb5a-5fcf5a44e2d5\") " pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.226041 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.270260 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 08:25:19 crc kubenswrapper[4948]: W0220 08:25:19.717811 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59965363_9971_4130_bb5a_5fcf5a44e2d5.slice/crio-76ceb8ccdbd76e943e8b1df093a8e84d25b146d383c80a19a3539cb59aa2e930 WatchSource:0}: Error finding container 76ceb8ccdbd76e943e8b1df093a8e84d25b146d383c80a19a3539cb59aa2e930: Status 404 returned error can't find the container with id 76ceb8ccdbd76e943e8b1df093a8e84d25b146d383c80a19a3539cb59aa2e930 Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.719186 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.762291 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61657c77-bb29-4241-9d88-5e13bccd3f4e" path="/var/lib/kubelet/pods/61657c77-bb29-4241-9d88-5e13bccd3f4e/volumes" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.762964 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="659ea49b-7e8f-4c95-ac69-ec743d0f95ed" path="/var/lib/kubelet/pods/659ea49b-7e8f-4c95-ac69-ec743d0f95ed/volumes" Feb 20 08:25:19 crc kubenswrapper[4948]: I0220 08:25:19.789995 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 08:25:19 crc kubenswrapper[4948]: W0220 08:25:19.801905 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod354f9b2f_9eaa_4e47_b214_0fd895f90c4c.slice/crio-5b932a3154a482569582f2a63f0553401470a58c072e8c1eab52792d5c807f8f WatchSource:0}: Error finding container 5b932a3154a482569582f2a63f0553401470a58c072e8c1eab52792d5c807f8f: Status 404 returned error can't find the container with id 5b932a3154a482569582f2a63f0553401470a58c072e8c1eab52792d5c807f8f Feb 20 08:25:20 crc kubenswrapper[4948]: I0220 08:25:20.569030 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"59965363-9971-4130-bb5a-5fcf5a44e2d5","Type":"ContainerStarted","Data":"6671f87799658c87c9825b2bf7885aedd1ec50455ebb820ad3145ff0cd4079bb"} Feb 20 08:25:20 crc kubenswrapper[4948]: I0220 08:25:20.569284 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"59965363-9971-4130-bb5a-5fcf5a44e2d5","Type":"ContainerStarted","Data":"c27f81fe70c91cfb56c95520ee8e8ab8efccada3d580de7a3beb5a071d8d9267"} Feb 20 08:25:20 crc kubenswrapper[4948]: I0220 08:25:20.569293 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"59965363-9971-4130-bb5a-5fcf5a44e2d5","Type":"ContainerStarted","Data":"76ceb8ccdbd76e943e8b1df093a8e84d25b146d383c80a19a3539cb59aa2e930"} Feb 20 08:25:20 crc kubenswrapper[4948]: I0220 08:25:20.572086 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"354f9b2f-9eaa-4e47-b214-0fd895f90c4c","Type":"ContainerStarted","Data":"fc3c49087b8c230b070d30b6ec36f8bd733f16769615b95321e2bb087eea68f1"} Feb 20 08:25:20 crc kubenswrapper[4948]: I0220 08:25:20.572115 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"354f9b2f-9eaa-4e47-b214-0fd895f90c4c","Type":"ContainerStarted","Data":"671fd39418c48112cbf952091c772751dc7a1df73ad2c05de77d69960fd1c5b5"} Feb 20 08:25:20 crc kubenswrapper[4948]: I0220 08:25:20.572124 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"354f9b2f-9eaa-4e47-b214-0fd895f90c4c","Type":"ContainerStarted","Data":"5b932a3154a482569582f2a63f0553401470a58c072e8c1eab52792d5c807f8f"} Feb 20 08:25:20 crc kubenswrapper[4948]: I0220 08:25:20.609573 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.609557508 podStartE2EDuration="2.609557508s" podCreationTimestamp="2026-02-20 08:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:25:20.604117086 +0000 UTC m=+1169.578611916" watchObservedRunningTime="2026-02-20 08:25:20.609557508 +0000 UTC m=+1169.584052318" Feb 20 08:25:20 crc kubenswrapper[4948]: I0220 08:25:20.639360 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.639344946 podStartE2EDuration="2.639344946s" podCreationTimestamp="2026-02-20 08:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:25:20.633591408 +0000 UTC m=+1169.608086228" watchObservedRunningTime="2026-02-20 08:25:20.639344946 +0000 UTC m=+1169.613839766" Feb 20 08:25:21 crc kubenswrapper[4948]: I0220 08:25:21.881303 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 20 08:25:24 crc kubenswrapper[4948]: I0220 08:25:24.226791 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 20 08:25:24 crc kubenswrapper[4948]: I0220 08:25:24.227378 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 20 08:25:26 crc kubenswrapper[4948]: I0220 08:25:26.881581 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 20 08:25:26 crc kubenswrapper[4948]: I0220 08:25:26.924538 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 20 08:25:27 crc kubenswrapper[4948]: I0220 08:25:27.693124 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 20 08:25:28 crc kubenswrapper[4948]: I0220 08:25:28.753356 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 20 08:25:29 crc kubenswrapper[4948]: I0220 08:25:29.226338 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 20 08:25:29 crc kubenswrapper[4948]: I0220 08:25:29.226676 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 20 08:25:29 crc kubenswrapper[4948]: I0220 08:25:29.271431 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 08:25:29 crc kubenswrapper[4948]: I0220 08:25:29.271517 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 08:25:30 crc kubenswrapper[4948]: I0220 08:25:30.241174 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="59965363-9971-4130-bb5a-5fcf5a44e2d5" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 08:25:30 crc kubenswrapper[4948]: I0220 08:25:30.241203 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="59965363-9971-4130-bb5a-5fcf5a44e2d5" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 08:25:30 crc kubenswrapper[4948]: I0220 08:25:30.287228 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="354f9b2f-9eaa-4e47-b214-0fd895f90c4c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 08:25:30 crc kubenswrapper[4948]: I0220 08:25:30.287277 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="354f9b2f-9eaa-4e47-b214-0fd895f90c4c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 08:25:32 crc kubenswrapper[4948]: I0220 08:25:32.854739 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:25:32 crc kubenswrapper[4948]: I0220 08:25:32.856467 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="0dfbb200-e444-4895-a410-e51cbb1fc112" containerName="kube-state-metrics" containerID="cri-o://20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0" gracePeriod=30 Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.348826 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.460746 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqgmb\" (UniqueName: \"kubernetes.io/projected/0dfbb200-e444-4895-a410-e51cbb1fc112-kube-api-access-rqgmb\") pod \"0dfbb200-e444-4895-a410-e51cbb1fc112\" (UID: \"0dfbb200-e444-4895-a410-e51cbb1fc112\") " Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.475265 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dfbb200-e444-4895-a410-e51cbb1fc112-kube-api-access-rqgmb" (OuterVolumeSpecName: "kube-api-access-rqgmb") pod "0dfbb200-e444-4895-a410-e51cbb1fc112" (UID: "0dfbb200-e444-4895-a410-e51cbb1fc112"). InnerVolumeSpecName "kube-api-access-rqgmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.563006 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqgmb\" (UniqueName: \"kubernetes.io/projected/0dfbb200-e444-4895-a410-e51cbb1fc112-kube-api-access-rqgmb\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.698311 4948 generic.go:334] "Generic (PLEG): container finished" podID="0dfbb200-e444-4895-a410-e51cbb1fc112" containerID="20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0" exitCode=2 Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.698354 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0dfbb200-e444-4895-a410-e51cbb1fc112","Type":"ContainerDied","Data":"20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0"} Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.698379 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0dfbb200-e444-4895-a410-e51cbb1fc112","Type":"ContainerDied","Data":"8d55ac91670695a37a460e364e6527e261fc751cd8cb9dd56d60bf1aa04e7514"} Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.698381 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.698406 4948 scope.go:117] "RemoveContainer" containerID="20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.729231 4948 scope.go:117] "RemoveContainer" containerID="20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0" Feb 20 08:25:33 crc kubenswrapper[4948]: E0220 08:25:33.729845 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0\": container with ID starting with 20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0 not found: ID does not exist" containerID="20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.729886 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0"} err="failed to get container status \"20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0\": rpc error: code = NotFound desc = could not find container \"20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0\": container with ID starting with 20c91950ce09f61ade4d935382ea88014ac2e6914fdd0f747d53e6577cf780a0 not found: ID does not exist" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.747625 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.769020 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.781656 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:25:33 crc kubenswrapper[4948]: E0220 08:25:33.782241 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dfbb200-e444-4895-a410-e51cbb1fc112" containerName="kube-state-metrics" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.782269 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dfbb200-e444-4895-a410-e51cbb1fc112" containerName="kube-state-metrics" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.782555 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dfbb200-e444-4895-a410-e51cbb1fc112" containerName="kube-state-metrics" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.783407 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.785153 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.788687 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.793581 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.868693 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdm4g\" (UniqueName: \"kubernetes.io/projected/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-api-access-sdm4g\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.868738 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.868951 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.869265 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.970792 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.970858 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.970935 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdm4g\" (UniqueName: \"kubernetes.io/projected/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-api-access-sdm4g\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.971334 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.974689 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.974823 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.975323 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:33 crc kubenswrapper[4948]: I0220 08:25:33.987772 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdm4g\" (UniqueName: \"kubernetes.io/projected/24bf0fb4-81db-4955-8b5f-92c0f6b0856b-kube-api-access-sdm4g\") pod \"kube-state-metrics-0\" (UID: \"24bf0fb4-81db-4955-8b5f-92c0f6b0856b\") " pod="openstack/kube-state-metrics-0" Feb 20 08:25:34 crc kubenswrapper[4948]: I0220 08:25:34.107278 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 08:25:34 crc kubenswrapper[4948]: I0220 08:25:34.626481 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 08:25:34 crc kubenswrapper[4948]: I0220 08:25:34.707687 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"24bf0fb4-81db-4955-8b5f-92c0f6b0856b","Type":"ContainerStarted","Data":"e2d46f0550c278b036f36f0b63c9d548764224188f6468a3d293b475c6b6027b"} Feb 20 08:25:34 crc kubenswrapper[4948]: I0220 08:25:34.788221 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:25:34 crc kubenswrapper[4948]: I0220 08:25:34.788658 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="proxy-httpd" containerID="cri-o://a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9" gracePeriod=30 Feb 20 08:25:34 crc kubenswrapper[4948]: I0220 08:25:34.788823 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="sg-core" containerID="cri-o://e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0" gracePeriod=30 Feb 20 08:25:34 crc kubenswrapper[4948]: I0220 08:25:34.788624 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="ceilometer-central-agent" containerID="cri-o://d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a" gracePeriod=30 Feb 20 08:25:34 crc kubenswrapper[4948]: I0220 08:25:34.789011 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="ceilometer-notification-agent" containerID="cri-o://43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542" gracePeriod=30 Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.736790 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"24bf0fb4-81db-4955-8b5f-92c0f6b0856b","Type":"ContainerStarted","Data":"8bbce6c2092a12dfd665eb6fc8e2c1e2aec7b889df488872597a31f777fdc8d3"} Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.748244 4948 generic.go:334] "Generic (PLEG): container finished" podID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerID="a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9" exitCode=0 Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.748274 4948 generic.go:334] "Generic (PLEG): container finished" podID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerID="e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0" exitCode=2 Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.748281 4948 generic.go:334] "Generic (PLEG): container finished" podID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerID="d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a" exitCode=0 Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.761951 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dfbb200-e444-4895-a410-e51cbb1fc112" path="/var/lib/kubelet/pods/0dfbb200-e444-4895-a410-e51cbb1fc112/volumes" Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.765435 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.765604 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerDied","Data":"a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9"} Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.765739 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerDied","Data":"e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0"} Feb 20 08:25:35 crc kubenswrapper[4948]: I0220 08:25:35.765855 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerDied","Data":"d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a"} Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.709094 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.738134 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.3084262349999998 podStartE2EDuration="3.738116545s" podCreationTimestamp="2026-02-20 08:25:33 +0000 UTC" firstStartedPulling="2026-02-20 08:25:34.611939256 +0000 UTC m=+1183.586434076" lastFinishedPulling="2026-02-20 08:25:35.041629566 +0000 UTC m=+1184.016124386" observedRunningTime="2026-02-20 08:25:35.761728813 +0000 UTC m=+1184.736223633" watchObservedRunningTime="2026-02-20 08:25:36.738116545 +0000 UTC m=+1185.712611365" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.762218 4948 generic.go:334] "Generic (PLEG): container finished" podID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerID="43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542" exitCode=0 Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.762281 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerDied","Data":"43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542"} Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.762336 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8","Type":"ContainerDied","Data":"48f0f09932c5f951969d8a7a57693456f1c18049262597223bdbf2f58c0804b2"} Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.762330 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.762365 4948 scope.go:117] "RemoveContainer" containerID="a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.787782 4948 scope.go:117] "RemoveContainer" containerID="e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.805247 4948 scope.go:117] "RemoveContainer" containerID="43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.824650 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-combined-ca-bundle\") pod \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.824809 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-run-httpd\") pod \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.824949 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-config-data\") pod \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.825048 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-sg-core-conf-yaml\") pod \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.825107 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-scripts\") pod \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.825129 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwwjr\" (UniqueName: \"kubernetes.io/projected/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-kube-api-access-mwwjr\") pod \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.825150 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-log-httpd\") pod \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\" (UID: \"be0e09bc-edd8-4ea9-8a59-ab630ed91cd8\") " Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.825222 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" (UID: "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.825567 4948 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.827518 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" (UID: "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.833331 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-scripts" (OuterVolumeSpecName: "scripts") pod "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" (UID: "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.833573 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-kube-api-access-mwwjr" (OuterVolumeSpecName: "kube-api-access-mwwjr") pod "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" (UID: "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8"). InnerVolumeSpecName "kube-api-access-mwwjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.848926 4948 scope.go:117] "RemoveContainer" containerID="d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.853109 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" (UID: "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.895847 4948 scope.go:117] "RemoveContainer" containerID="a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9" Feb 20 08:25:36 crc kubenswrapper[4948]: E0220 08:25:36.900517 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9\": container with ID starting with a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9 not found: ID does not exist" containerID="a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.900548 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9"} err="failed to get container status \"a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9\": rpc error: code = NotFound desc = could not find container \"a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9\": container with ID starting with a1bbb73bd50190b5b08c60773344adf5c06efb89abf70ae7646fd73534b398b9 not found: ID does not exist" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.900568 4948 scope.go:117] "RemoveContainer" containerID="e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0" Feb 20 08:25:36 crc kubenswrapper[4948]: E0220 08:25:36.902523 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0\": container with ID starting with e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0 not found: ID does not exist" containerID="e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.902553 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0"} err="failed to get container status \"e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0\": rpc error: code = NotFound desc = could not find container \"e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0\": container with ID starting with e04f1d97ea889929a68ccc1a2406ae3771d7b5ea9af14caaec52d943991404e0 not found: ID does not exist" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.902570 4948 scope.go:117] "RemoveContainer" containerID="43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542" Feb 20 08:25:36 crc kubenswrapper[4948]: E0220 08:25:36.903018 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542\": container with ID starting with 43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542 not found: ID does not exist" containerID="43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.903045 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542"} err="failed to get container status \"43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542\": rpc error: code = NotFound desc = could not find container \"43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542\": container with ID starting with 43e2a82447778e081804b3e50b6f485e50305ec4eb1bbe2f231f259d11b87542 not found: ID does not exist" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.903061 4948 scope.go:117] "RemoveContainer" containerID="d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a" Feb 20 08:25:36 crc kubenswrapper[4948]: E0220 08:25:36.903387 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a\": container with ID starting with d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a not found: ID does not exist" containerID="d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.903428 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a"} err="failed to get container status \"d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a\": rpc error: code = NotFound desc = could not find container \"d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a\": container with ID starting with d03b827c5401b1d3755870ce6ce51a7201888c77eda8df6a4497ae437d66e88a not found: ID does not exist" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.927011 4948 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.927036 4948 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.927046 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwwjr\" (UniqueName: \"kubernetes.io/projected/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-kube-api-access-mwwjr\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.927054 4948 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.936146 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-config-data" (OuterVolumeSpecName: "config-data") pod "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" (UID: "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:36 crc kubenswrapper[4948]: I0220 08:25:36.946121 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" (UID: "be0e09bc-edd8-4ea9-8a59-ab630ed91cd8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.029088 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.029124 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.097821 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.106637 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.129156 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:25:37 crc kubenswrapper[4948]: E0220 08:25:37.129611 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="proxy-httpd" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.129635 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="proxy-httpd" Feb 20 08:25:37 crc kubenswrapper[4948]: E0220 08:25:37.129666 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="sg-core" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.129676 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="sg-core" Feb 20 08:25:37 crc kubenswrapper[4948]: E0220 08:25:37.129684 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="ceilometer-notification-agent" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.129692 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="ceilometer-notification-agent" Feb 20 08:25:37 crc kubenswrapper[4948]: E0220 08:25:37.129723 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="ceilometer-central-agent" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.129731 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="ceilometer-central-agent" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.129930 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="ceilometer-notification-agent" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.129953 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="sg-core" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.129991 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="ceilometer-central-agent" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.130012 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" containerName="proxy-httpd" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.132039 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.136919 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.137783 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.138047 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.149200 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.231734 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58f563bb-9716-4abd-b187-3f01036d4f31-log-httpd\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.231818 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58f563bb-9716-4abd-b187-3f01036d4f31-run-httpd\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.231851 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.231888 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-config-data\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.231907 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.231936 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj8mt\" (UniqueName: \"kubernetes.io/projected/58f563bb-9716-4abd-b187-3f01036d4f31-kube-api-access-xj8mt\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.232004 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.232031 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-scripts\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.333603 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-scripts\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.333733 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58f563bb-9716-4abd-b187-3f01036d4f31-log-httpd\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.333785 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58f563bb-9716-4abd-b187-3f01036d4f31-run-httpd\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.333825 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.333871 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-config-data\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.333892 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.333922 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj8mt\" (UniqueName: \"kubernetes.io/projected/58f563bb-9716-4abd-b187-3f01036d4f31-kube-api-access-xj8mt\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.333999 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.335057 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58f563bb-9716-4abd-b187-3f01036d4f31-run-httpd\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.335637 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58f563bb-9716-4abd-b187-3f01036d4f31-log-httpd\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.339000 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.341666 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.343083 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-scripts\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.347377 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.348839 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58f563bb-9716-4abd-b187-3f01036d4f31-config-data\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.362243 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj8mt\" (UniqueName: \"kubernetes.io/projected/58f563bb-9716-4abd-b187-3f01036d4f31-kube-api-access-xj8mt\") pod \"ceilometer-0\" (UID: \"58f563bb-9716-4abd-b187-3f01036d4f31\") " pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.445818 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.733512 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be0e09bc-edd8-4ea9-8a59-ab630ed91cd8" path="/var/lib/kubelet/pods/be0e09bc-edd8-4ea9-8a59-ab630ed91cd8/volumes" Feb 20 08:25:37 crc kubenswrapper[4948]: I0220 08:25:37.809873 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 08:25:37 crc kubenswrapper[4948]: W0220 08:25:37.814395 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58f563bb_9716_4abd_b187_3f01036d4f31.slice/crio-f2e44f0bba52f19eac0a54be4d80b29f6c40e12f4cd9b659958bfeb7f47736f8 WatchSource:0}: Error finding container f2e44f0bba52f19eac0a54be4d80b29f6c40e12f4cd9b659958bfeb7f47736f8: Status 404 returned error can't find the container with id f2e44f0bba52f19eac0a54be4d80b29f6c40e12f4cd9b659958bfeb7f47736f8 Feb 20 08:25:38 crc kubenswrapper[4948]: I0220 08:25:38.800034 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58f563bb-9716-4abd-b187-3f01036d4f31","Type":"ContainerStarted","Data":"48565919319b61cf40009b1fc3e376ca0610ca1b8cffb64ba96185c28a75764b"} Feb 20 08:25:38 crc kubenswrapper[4948]: I0220 08:25:38.800446 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58f563bb-9716-4abd-b187-3f01036d4f31","Type":"ContainerStarted","Data":"f2e44f0bba52f19eac0a54be4d80b29f6c40e12f4cd9b659958bfeb7f47736f8"} Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.231620 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.233216 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.240355 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.281490 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.285168 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.292165 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.300233 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.812493 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58f563bb-9716-4abd-b187-3f01036d4f31","Type":"ContainerStarted","Data":"9d613f1e4dac4655d24f385e90b54a96a440e203bbaebef9c0b52d894937a29e"} Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.813233 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.816535 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 20 08:25:39 crc kubenswrapper[4948]: I0220 08:25:39.820938 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 20 08:25:40 crc kubenswrapper[4948]: I0220 08:25:40.824385 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58f563bb-9716-4abd-b187-3f01036d4f31","Type":"ContainerStarted","Data":"94b2bd264871b0e4898e842ab549234e35ff3a87346db7a895e5a9b2b8fc5897"} Feb 20 08:25:41 crc kubenswrapper[4948]: I0220 08:25:41.836459 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58f563bb-9716-4abd-b187-3f01036d4f31","Type":"ContainerStarted","Data":"3bcc9d21f8f1da159817d123e65bc83691e3ce0ff5eff3f6c34f4a99d28e9433"} Feb 20 08:25:41 crc kubenswrapper[4948]: I0220 08:25:41.867891 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.321293442 podStartE2EDuration="4.867864228s" podCreationTimestamp="2026-02-20 08:25:37 +0000 UTC" firstStartedPulling="2026-02-20 08:25:37.817180576 +0000 UTC m=+1186.791675416" lastFinishedPulling="2026-02-20 08:25:41.363751382 +0000 UTC m=+1190.338246202" observedRunningTime="2026-02-20 08:25:41.85634308 +0000 UTC m=+1190.830837940" watchObservedRunningTime="2026-02-20 08:25:41.867864228 +0000 UTC m=+1190.842359088" Feb 20 08:25:42 crc kubenswrapper[4948]: I0220 08:25:42.854258 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 08:25:44 crc kubenswrapper[4948]: I0220 08:25:44.134112 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 20 08:26:07 crc kubenswrapper[4948]: I0220 08:26:07.454550 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 20 08:26:16 crc kubenswrapper[4948]: I0220 08:26:16.577316 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:26:17 crc kubenswrapper[4948]: I0220 08:26:17.598386 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:26:20 crc kubenswrapper[4948]: I0220 08:26:20.589002 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerName="rabbitmq" containerID="cri-o://f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85" gracePeriod=604796 Feb 20 08:26:21 crc kubenswrapper[4948]: I0220 08:26:21.523688 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="790894c9-053c-497c-955d-ce7519111dd6" containerName="rabbitmq" containerID="cri-o://ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a" gracePeriod=604797 Feb 20 08:26:24 crc kubenswrapper[4948]: I0220 08:26:24.117591 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.94:5671: connect: connection refused" Feb 20 08:26:24 crc kubenswrapper[4948]: I0220 08:26:24.498402 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="790894c9-053c-497c-955d-ce7519111dd6" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.95:5671: connect: connection refused" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.289632 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349154 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/73b84bb7-f594-4823-ac03-40fdac6ee177-pod-info\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349228 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-server-conf\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349244 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-tls\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349270 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-config-data\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349301 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-erlang-cookie\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349347 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349380 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/73b84bb7-f594-4823-ac03-40fdac6ee177-erlang-cookie-secret\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349428 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-plugins\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349471 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-plugins-conf\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349490 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-confd\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.349556 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqlcq\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-kube-api-access-tqlcq\") pod \"73b84bb7-f594-4823-ac03-40fdac6ee177\" (UID: \"73b84bb7-f594-4823-ac03-40fdac6ee177\") " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.350426 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.350628 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.350733 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.356708 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-kube-api-access-tqlcq" (OuterVolumeSpecName: "kube-api-access-tqlcq") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "kube-api-access-tqlcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.358193 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/73b84bb7-f594-4823-ac03-40fdac6ee177-pod-info" (OuterVolumeSpecName: "pod-info") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.358424 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73b84bb7-f594-4823-ac03-40fdac6ee177-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.367317 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.374196 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.415847 4948 generic.go:334] "Generic (PLEG): container finished" podID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerID="f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85" exitCode=0 Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.415891 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"73b84bb7-f594-4823-ac03-40fdac6ee177","Type":"ContainerDied","Data":"f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85"} Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.415916 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"73b84bb7-f594-4823-ac03-40fdac6ee177","Type":"ContainerDied","Data":"e552b8693d7f82c1ed5a2c7c103ec119814c444b37d5f93e2491692378454a4c"} Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.415932 4948 scope.go:117] "RemoveContainer" containerID="f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.416796 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.427239 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-config-data" (OuterVolumeSpecName: "config-data") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.452577 4948 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/73b84bb7-f594-4823-ac03-40fdac6ee177-pod-info\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.452609 4948 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.452624 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.452636 4948 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.452682 4948 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.452694 4948 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/73b84bb7-f594-4823-ac03-40fdac6ee177-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.452708 4948 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.459045 4948 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.459095 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqlcq\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-kube-api-access-tqlcq\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.476131 4948 scope.go:117] "RemoveContainer" containerID="d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.477161 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-server-conf" (OuterVolumeSpecName: "server-conf") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.501147 4948 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.514873 4948 scope.go:117] "RemoveContainer" containerID="f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85" Feb 20 08:26:27 crc kubenswrapper[4948]: E0220 08:26:27.515316 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85\": container with ID starting with f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85 not found: ID does not exist" containerID="f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.515344 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85"} err="failed to get container status \"f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85\": rpc error: code = NotFound desc = could not find container \"f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85\": container with ID starting with f2ef52eaf8fede6026f04fb9c9c143db4ac26c52cfdab80a2a52811b838b1e85 not found: ID does not exist" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.515362 4948 scope.go:117] "RemoveContainer" containerID="d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364" Feb 20 08:26:27 crc kubenswrapper[4948]: E0220 08:26:27.515696 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364\": container with ID starting with d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364 not found: ID does not exist" containerID="d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.515716 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364"} err="failed to get container status \"d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364\": rpc error: code = NotFound desc = could not find container \"d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364\": container with ID starting with d4b1f06d6e332e362b758285a5271800ff6a5cc95aee58142ef4093fc832e364 not found: ID does not exist" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.551504 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "73b84bb7-f594-4823-ac03-40fdac6ee177" (UID: "73b84bb7-f594-4823-ac03-40fdac6ee177"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.561620 4948 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/73b84bb7-f594-4823-ac03-40fdac6ee177-server-conf\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.562009 4948 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.562083 4948 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/73b84bb7-f594-4823-ac03-40fdac6ee177-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.766423 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.787412 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.804985 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:26:27 crc kubenswrapper[4948]: E0220 08:26:27.805629 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerName="rabbitmq" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.805768 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerName="rabbitmq" Feb 20 08:26:27 crc kubenswrapper[4948]: E0220 08:26:27.805860 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerName="setup-container" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.805941 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerName="setup-container" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.806267 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="73b84bb7-f594-4823-ac03-40fdac6ee177" containerName="rabbitmq" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.807649 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.811538 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.811834 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.826253 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.827283 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.827486 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.827598 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.827696 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.827795 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-8zcmc" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868096 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f063bb62-10c3-46d9-a41d-56d2826059e0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868141 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjqdv\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-kube-api-access-mjqdv\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868163 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868201 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f063bb62-10c3-46d9-a41d-56d2826059e0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868225 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868241 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868331 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868371 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868411 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868441 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-config-data\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.868482 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970191 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970274 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970328 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-config-data\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970388 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970427 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f063bb62-10c3-46d9-a41d-56d2826059e0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970467 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjqdv\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-kube-api-access-mjqdv\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970485 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970513 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f063bb62-10c3-46d9-a41d-56d2826059e0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970543 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.972137 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.970553 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.978485 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.978577 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.979549 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-config-data\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.980052 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.980416 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.980520 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f063bb62-10c3-46d9-a41d-56d2826059e0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.983102 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.985995 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f063bb62-10c3-46d9-a41d-56d2826059e0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.992329 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:27 crc kubenswrapper[4948]: I0220 08:26:27.995558 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f063bb62-10c3-46d9-a41d-56d2826059e0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.002144 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjqdv\" (UniqueName: \"kubernetes.io/projected/f063bb62-10c3-46d9-a41d-56d2826059e0-kube-api-access-mjqdv\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.032669 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"f063bb62-10c3-46d9-a41d-56d2826059e0\") " pod="openstack/rabbitmq-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.127685 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.193813 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.282662 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-plugins-conf\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.282894 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-plugins\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283024 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7t2f\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-kube-api-access-b7t2f\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283085 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-tls\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283113 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/790894c9-053c-497c-955d-ce7519111dd6-erlang-cookie-secret\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283164 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-confd\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283211 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-server-conf\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283235 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283277 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/790894c9-053c-497c-955d-ce7519111dd6-pod-info\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283309 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-erlang-cookie\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283338 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-config-data\") pod \"790894c9-053c-497c-955d-ce7519111dd6\" (UID: \"790894c9-053c-497c-955d-ce7519111dd6\") " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283548 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283863 4948 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.283863 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.289690 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.295770 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.295785 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/790894c9-053c-497c-955d-ce7519111dd6-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.296304 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.298357 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-kube-api-access-b7t2f" (OuterVolumeSpecName: "kube-api-access-b7t2f") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "kube-api-access-b7t2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.299100 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/790894c9-053c-497c-955d-ce7519111dd6-pod-info" (OuterVolumeSpecName: "pod-info") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.343736 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-config-data" (OuterVolumeSpecName: "config-data") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.356996 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-server-conf" (OuterVolumeSpecName: "server-conf") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385712 4948 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385742 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7t2f\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-kube-api-access-b7t2f\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385752 4948 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385763 4948 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/790894c9-053c-497c-955d-ce7519111dd6-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385771 4948 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-server-conf\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385790 4948 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385800 4948 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/790894c9-053c-497c-955d-ce7519111dd6-pod-info\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385808 4948 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.385816 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/790894c9-053c-497c-955d-ce7519111dd6-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.409318 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "790894c9-053c-497c-955d-ce7519111dd6" (UID: "790894c9-053c-497c-955d-ce7519111dd6"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.415366 4948 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.425897 4948 generic.go:334] "Generic (PLEG): container finished" podID="790894c9-053c-497c-955d-ce7519111dd6" containerID="ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a" exitCode=0 Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.425961 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"790894c9-053c-497c-955d-ce7519111dd6","Type":"ContainerDied","Data":"ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a"} Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.426001 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.426065 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"790894c9-053c-497c-955d-ce7519111dd6","Type":"ContainerDied","Data":"39af94a050f39ff7ebfa915fc855416f40bbca9d8ae91ba23049ec9be5e37b76"} Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.426103 4948 scope.go:117] "RemoveContainer" containerID="ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.458416 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.472054 4948 scope.go:117] "RemoveContainer" containerID="77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.478122 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.487261 4948 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/790894c9-053c-497c-955d-ce7519111dd6-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.487290 4948 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.494818 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:26:28 crc kubenswrapper[4948]: E0220 08:26:28.495207 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790894c9-053c-497c-955d-ce7519111dd6" containerName="setup-container" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.495227 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="790894c9-053c-497c-955d-ce7519111dd6" containerName="setup-container" Feb 20 08:26:28 crc kubenswrapper[4948]: E0220 08:26:28.495252 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790894c9-053c-497c-955d-ce7519111dd6" containerName="rabbitmq" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.495259 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="790894c9-053c-497c-955d-ce7519111dd6" containerName="rabbitmq" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.495439 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="790894c9-053c-497c-955d-ce7519111dd6" containerName="rabbitmq" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.496303 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.501695 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.501713 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.501875 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.502299 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.502395 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.502476 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-rrjt5" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.502919 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.515070 4948 scope.go:117] "RemoveContainer" containerID="ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.515585 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:26:28 crc kubenswrapper[4948]: E0220 08:26:28.520793 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a\": container with ID starting with ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a not found: ID does not exist" containerID="ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.520827 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a"} err="failed to get container status \"ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a\": rpc error: code = NotFound desc = could not find container \"ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a\": container with ID starting with ba12ca5de96a23b6d0de36b28eae9e3fb2a92d4464edd4582c10c73ae804254a not found: ID does not exist" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.520848 4948 scope.go:117] "RemoveContainer" containerID="77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242" Feb 20 08:26:28 crc kubenswrapper[4948]: E0220 08:26:28.527463 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242\": container with ID starting with 77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242 not found: ID does not exist" containerID="77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.527650 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242"} err="failed to get container status \"77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242\": rpc error: code = NotFound desc = could not find container \"77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242\": container with ID starting with 77b026b61409d73dee6a89b7f49bf1a86167df7972db0e3eb3f4cbb10df48242 not found: ID does not exist" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588289 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588341 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588383 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588425 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn7kw\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-kube-api-access-fn7kw\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588452 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588471 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588494 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588514 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588559 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588573 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/037b329f-712f-4a67-984a-75affd2a57b7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.588594 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/037b329f-712f-4a67-984a-75affd2a57b7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.651421 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690347 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690399 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/037b329f-712f-4a67-984a-75affd2a57b7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690440 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/037b329f-712f-4a67-984a-75affd2a57b7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690463 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690499 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690549 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690599 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn7kw\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-kube-api-access-fn7kw\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690635 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690663 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690726 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.690759 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.691297 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.691437 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.692711 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.693318 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.693530 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/037b329f-712f-4a67-984a-75affd2a57b7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.693667 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.695055 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.696171 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/037b329f-712f-4a67-984a-75affd2a57b7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.700467 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/037b329f-712f-4a67-984a-75affd2a57b7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.707460 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.710169 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn7kw\" (UniqueName: \"kubernetes.io/projected/037b329f-712f-4a67-984a-75affd2a57b7-kube-api-access-fn7kw\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.736956 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"037b329f-712f-4a67-984a-75affd2a57b7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:28 crc kubenswrapper[4948]: I0220 08:26:28.822113 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.166655 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-vgk45"] Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.169046 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.171304 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.193784 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-vgk45"] Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.200355 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.200421 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.200465 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-config\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.200565 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.200614 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.200658 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.200726 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4h8d\" (UniqueName: \"kubernetes.io/projected/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-kube-api-access-x4h8d\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.302075 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.302149 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4h8d\" (UniqueName: \"kubernetes.io/projected/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-kube-api-access-x4h8d\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.302194 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.302219 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.302256 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-config\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.302317 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.302349 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.303354 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.303881 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.304634 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.305259 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.305792 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-config\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.306018 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.306245 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.322835 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4h8d\" (UniqueName: \"kubernetes.io/projected/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-kube-api-access-x4h8d\") pod \"dnsmasq-dns-79bd4cc8c9-vgk45\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.444335 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"037b329f-712f-4a67-984a-75affd2a57b7","Type":"ContainerStarted","Data":"5523f651b79457ea6012384e30f4ce29cd629dcb583a6414256e8cccebe13e53"} Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.446479 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f063bb62-10c3-46d9-a41d-56d2826059e0","Type":"ContainerStarted","Data":"49966811dc211199039188fd9f5a17caac95e7a7a6cc21002f62432ed349328d"} Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.495117 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.752025 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73b84bb7-f594-4823-ac03-40fdac6ee177" path="/var/lib/kubelet/pods/73b84bb7-f594-4823-ac03-40fdac6ee177/volumes" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.754326 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="790894c9-053c-497c-955d-ce7519111dd6" path="/var/lib/kubelet/pods/790894c9-053c-497c-955d-ce7519111dd6/volumes" Feb 20 08:26:29 crc kubenswrapper[4948]: I0220 08:26:29.998387 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-vgk45"] Feb 20 08:26:29 crc kubenswrapper[4948]: W0220 08:26:29.999620 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4bae750_8bf9_4804_bab7_9c94d4be7bb6.slice/crio-a5fc6b8697916867d170463ef9a68553f0374e86750fe3d5b67599ef15a68b91 WatchSource:0}: Error finding container a5fc6b8697916867d170463ef9a68553f0374e86750fe3d5b67599ef15a68b91: Status 404 returned error can't find the container with id a5fc6b8697916867d170463ef9a68553f0374e86750fe3d5b67599ef15a68b91 Feb 20 08:26:30 crc kubenswrapper[4948]: I0220 08:26:30.466074 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f063bb62-10c3-46d9-a41d-56d2826059e0","Type":"ContainerStarted","Data":"878f9968270125c43cab5cdbc68ca06efa6e152bfb9c605c96475afc50f8ad38"} Feb 20 08:26:30 crc kubenswrapper[4948]: I0220 08:26:30.470427 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" event={"ID":"b4bae750-8bf9-4804-bab7-9c94d4be7bb6","Type":"ContainerStarted","Data":"a5fc6b8697916867d170463ef9a68553f0374e86750fe3d5b67599ef15a68b91"} Feb 20 08:26:31 crc kubenswrapper[4948]: I0220 08:26:31.482959 4948 generic.go:334] "Generic (PLEG): container finished" podID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" containerID="6457842f65c42b8313a38ca922ee61a97a4823ed4c56040981289a5c5705cd46" exitCode=0 Feb 20 08:26:31 crc kubenswrapper[4948]: I0220 08:26:31.483042 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" event={"ID":"b4bae750-8bf9-4804-bab7-9c94d4be7bb6","Type":"ContainerDied","Data":"6457842f65c42b8313a38ca922ee61a97a4823ed4c56040981289a5c5705cd46"} Feb 20 08:26:31 crc kubenswrapper[4948]: I0220 08:26:31.485969 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"037b329f-712f-4a67-984a-75affd2a57b7","Type":"ContainerStarted","Data":"9fcd3d9744d41147b3cfe90374517e1e56f49253e12cf9785df1944e69f8e3a2"} Feb 20 08:26:32 crc kubenswrapper[4948]: I0220 08:26:32.506872 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" event={"ID":"b4bae750-8bf9-4804-bab7-9c94d4be7bb6","Type":"ContainerStarted","Data":"ede14cf7eee35beb81bccaf96c5abbd270f76ba98b65d67579bbb482b130f45e"} Feb 20 08:26:32 crc kubenswrapper[4948]: I0220 08:26:32.507355 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:32 crc kubenswrapper[4948]: I0220 08:26:32.545562 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" podStartSLOduration=3.545532165 podStartE2EDuration="3.545532165s" podCreationTimestamp="2026-02-20 08:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:26:32.536469842 +0000 UTC m=+1241.510964702" watchObservedRunningTime="2026-02-20 08:26:32.545532165 +0000 UTC m=+1241.520027015" Feb 20 08:26:38 crc kubenswrapper[4948]: I0220 08:26:38.024811 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:26:38 crc kubenswrapper[4948]: I0220 08:26:38.025429 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.497503 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.609662 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-t56xb"] Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.609948 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" podUID="c1847849-cdf4-4ea6-a540-c01503a003b2" containerName="dnsmasq-dns" containerID="cri-o://5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f" gracePeriod=10 Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.805445 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55478c4467-fbjpk"] Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.811756 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.820757 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-fbjpk"] Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.831595 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.831638 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.831688 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.831750 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xksr2\" (UniqueName: \"kubernetes.io/projected/b05f92c6-60c9-41be-bf6c-6288cee04659-kube-api-access-xksr2\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.831793 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-config\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.831826 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-dns-svc\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.831899 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.933429 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.933483 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.933558 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.933617 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xksr2\" (UniqueName: \"kubernetes.io/projected/b05f92c6-60c9-41be-bf6c-6288cee04659-kube-api-access-xksr2\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.933667 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-config\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.933714 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-dns-svc\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.933737 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.934783 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.934844 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.934869 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-dns-svc\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.934890 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-config\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.936874 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.936907 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b05f92c6-60c9-41be-bf6c-6288cee04659-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:39 crc kubenswrapper[4948]: I0220 08:26:39.967132 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xksr2\" (UniqueName: \"kubernetes.io/projected/b05f92c6-60c9-41be-bf6c-6288cee04659-kube-api-access-xksr2\") pod \"dnsmasq-dns-55478c4467-fbjpk\" (UID: \"b05f92c6-60c9-41be-bf6c-6288cee04659\") " pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.130019 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.145345 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.342071 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-nb\") pod \"c1847849-cdf4-4ea6-a540-c01503a003b2\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.342500 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-svc\") pod \"c1847849-cdf4-4ea6-a540-c01503a003b2\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.342551 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-config\") pod \"c1847849-cdf4-4ea6-a540-c01503a003b2\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.342703 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-db22s\" (UniqueName: \"kubernetes.io/projected/c1847849-cdf4-4ea6-a540-c01503a003b2-kube-api-access-db22s\") pod \"c1847849-cdf4-4ea6-a540-c01503a003b2\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.342739 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-sb\") pod \"c1847849-cdf4-4ea6-a540-c01503a003b2\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.343519 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-swift-storage-0\") pod \"c1847849-cdf4-4ea6-a540-c01503a003b2\" (UID: \"c1847849-cdf4-4ea6-a540-c01503a003b2\") " Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.349092 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1847849-cdf4-4ea6-a540-c01503a003b2-kube-api-access-db22s" (OuterVolumeSpecName: "kube-api-access-db22s") pod "c1847849-cdf4-4ea6-a540-c01503a003b2" (UID: "c1847849-cdf4-4ea6-a540-c01503a003b2"). InnerVolumeSpecName "kube-api-access-db22s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.405292 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c1847849-cdf4-4ea6-a540-c01503a003b2" (UID: "c1847849-cdf4-4ea6-a540-c01503a003b2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.405571 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c1847849-cdf4-4ea6-a540-c01503a003b2" (UID: "c1847849-cdf4-4ea6-a540-c01503a003b2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.409160 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-config" (OuterVolumeSpecName: "config") pod "c1847849-cdf4-4ea6-a540-c01503a003b2" (UID: "c1847849-cdf4-4ea6-a540-c01503a003b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.413766 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c1847849-cdf4-4ea6-a540-c01503a003b2" (UID: "c1847849-cdf4-4ea6-a540-c01503a003b2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.427498 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c1847849-cdf4-4ea6-a540-c01503a003b2" (UID: "c1847849-cdf4-4ea6-a540-c01503a003b2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.445637 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.445684 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.445698 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-db22s\" (UniqueName: \"kubernetes.io/projected/c1847849-cdf4-4ea6-a540-c01503a003b2-kube-api-access-db22s\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.445713 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.445725 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.445736 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1847849-cdf4-4ea6-a540-c01503a003b2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.594191 4948 generic.go:334] "Generic (PLEG): container finished" podID="c1847849-cdf4-4ea6-a540-c01503a003b2" containerID="5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f" exitCode=0 Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.594238 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.594250 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" event={"ID":"c1847849-cdf4-4ea6-a540-c01503a003b2","Type":"ContainerDied","Data":"5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f"} Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.594293 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-t56xb" event={"ID":"c1847849-cdf4-4ea6-a540-c01503a003b2","Type":"ContainerDied","Data":"4b70d018879f9b501eed3beab4a8c0781964989ca45e86dd4327f865ae78f6bf"} Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.594315 4948 scope.go:117] "RemoveContainer" containerID="5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.597965 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-fbjpk"] Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.631714 4948 scope.go:117] "RemoveContainer" containerID="6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.634751 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-t56xb"] Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.645136 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-t56xb"] Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.666425 4948 scope.go:117] "RemoveContainer" containerID="5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f" Feb 20 08:26:40 crc kubenswrapper[4948]: E0220 08:26:40.666876 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f\": container with ID starting with 5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f not found: ID does not exist" containerID="5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.667000 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f"} err="failed to get container status \"5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f\": rpc error: code = NotFound desc = could not find container \"5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f\": container with ID starting with 5c65f781c2289e22368c8df4db43181d77d0c56ecf63f298c154f0b8af27a14f not found: ID does not exist" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.667090 4948 scope.go:117] "RemoveContainer" containerID="6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a" Feb 20 08:26:40 crc kubenswrapper[4948]: E0220 08:26:40.667540 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a\": container with ID starting with 6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a not found: ID does not exist" containerID="6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a" Feb 20 08:26:40 crc kubenswrapper[4948]: I0220 08:26:40.668170 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a"} err="failed to get container status \"6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a\": rpc error: code = NotFound desc = could not find container \"6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a\": container with ID starting with 6b70c340bd4972b75cf8aa94cef8eb5d658d1c6345ae3959074d416c3683e47a not found: ID does not exist" Feb 20 08:26:41 crc kubenswrapper[4948]: I0220 08:26:41.605352 4948 generic.go:334] "Generic (PLEG): container finished" podID="b05f92c6-60c9-41be-bf6c-6288cee04659" containerID="bdea951a4170c0cdceb0ddcae984666862ac3f74ed32191e1024d762cf306aa7" exitCode=0 Feb 20 08:26:41 crc kubenswrapper[4948]: I0220 08:26:41.605451 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-fbjpk" event={"ID":"b05f92c6-60c9-41be-bf6c-6288cee04659","Type":"ContainerDied","Data":"bdea951a4170c0cdceb0ddcae984666862ac3f74ed32191e1024d762cf306aa7"} Feb 20 08:26:41 crc kubenswrapper[4948]: I0220 08:26:41.607682 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-fbjpk" event={"ID":"b05f92c6-60c9-41be-bf6c-6288cee04659","Type":"ContainerStarted","Data":"8881488afcebfb5387b7d34be5b2ee2caf4be9af99e858b4f8b197f620f5b16f"} Feb 20 08:26:41 crc kubenswrapper[4948]: I0220 08:26:41.735406 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1847849-cdf4-4ea6-a540-c01503a003b2" path="/var/lib/kubelet/pods/c1847849-cdf4-4ea6-a540-c01503a003b2/volumes" Feb 20 08:26:42 crc kubenswrapper[4948]: I0220 08:26:42.630865 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-fbjpk" event={"ID":"b05f92c6-60c9-41be-bf6c-6288cee04659","Type":"ContainerStarted","Data":"180d9e61aab4c578242860dafa196fb5407f99775735708fe0b437e89f49320a"} Feb 20 08:26:42 crc kubenswrapper[4948]: I0220 08:26:42.631060 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:42 crc kubenswrapper[4948]: I0220 08:26:42.679102 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55478c4467-fbjpk" podStartSLOduration=3.679075083 podStartE2EDuration="3.679075083s" podCreationTimestamp="2026-02-20 08:26:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:26:42.667081947 +0000 UTC m=+1251.641576837" watchObservedRunningTime="2026-02-20 08:26:42.679075083 +0000 UTC m=+1251.653569933" Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.133237 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55478c4467-fbjpk" Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.229745 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-vgk45"] Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.231284 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" podUID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" containerName="dnsmasq-dns" containerID="cri-o://ede14cf7eee35beb81bccaf96c5abbd270f76ba98b65d67579bbb482b130f45e" gracePeriod=10 Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.729353 4948 generic.go:334] "Generic (PLEG): container finished" podID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" containerID="ede14cf7eee35beb81bccaf96c5abbd270f76ba98b65d67579bbb482b130f45e" exitCode=0 Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.729474 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" event={"ID":"b4bae750-8bf9-4804-bab7-9c94d4be7bb6","Type":"ContainerDied","Data":"ede14cf7eee35beb81bccaf96c5abbd270f76ba98b65d67579bbb482b130f45e"} Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.729710 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" event={"ID":"b4bae750-8bf9-4804-bab7-9c94d4be7bb6","Type":"ContainerDied","Data":"a5fc6b8697916867d170463ef9a68553f0374e86750fe3d5b67599ef15a68b91"} Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.729726 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5fc6b8697916867d170463ef9a68553f0374e86750fe3d5b67599ef15a68b91" Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.795573 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.980535 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-config\") pod \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.980608 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-nb\") pod \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.980643 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-svc\") pod \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.980679 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-sb\") pod \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.980710 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-swift-storage-0\") pod \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.980831 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4h8d\" (UniqueName: \"kubernetes.io/projected/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-kube-api-access-x4h8d\") pod \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " Feb 20 08:26:50 crc kubenswrapper[4948]: I0220 08:26:50.980873 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-openstack-edpm-ipam\") pod \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\" (UID: \"b4bae750-8bf9-4804-bab7-9c94d4be7bb6\") " Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.010171 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-kube-api-access-x4h8d" (OuterVolumeSpecName: "kube-api-access-x4h8d") pod "b4bae750-8bf9-4804-bab7-9c94d4be7bb6" (UID: "b4bae750-8bf9-4804-bab7-9c94d4be7bb6"). InnerVolumeSpecName "kube-api-access-x4h8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.084034 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4h8d\" (UniqueName: \"kubernetes.io/projected/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-kube-api-access-x4h8d\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.104496 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b4bae750-8bf9-4804-bab7-9c94d4be7bb6" (UID: "b4bae750-8bf9-4804-bab7-9c94d4be7bb6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.110342 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b4bae750-8bf9-4804-bab7-9c94d4be7bb6" (UID: "b4bae750-8bf9-4804-bab7-9c94d4be7bb6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.116793 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "b4bae750-8bf9-4804-bab7-9c94d4be7bb6" (UID: "b4bae750-8bf9-4804-bab7-9c94d4be7bb6"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.126572 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b4bae750-8bf9-4804-bab7-9c94d4be7bb6" (UID: "b4bae750-8bf9-4804-bab7-9c94d4be7bb6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.142795 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b4bae750-8bf9-4804-bab7-9c94d4be7bb6" (UID: "b4bae750-8bf9-4804-bab7-9c94d4be7bb6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.147572 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-config" (OuterVolumeSpecName: "config") pod "b4bae750-8bf9-4804-bab7-9c94d4be7bb6" (UID: "b4bae750-8bf9-4804-bab7-9c94d4be7bb6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.185811 4948 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.185860 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.185869 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.185879 4948 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.185889 4948 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.185897 4948 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4bae750-8bf9-4804-bab7-9c94d4be7bb6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.740366 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-vgk45" Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.804213 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-vgk45"] Feb 20 08:26:51 crc kubenswrapper[4948]: I0220 08:26:51.815682 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-vgk45"] Feb 20 08:26:53 crc kubenswrapper[4948]: I0220 08:26:53.737192 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" path="/var/lib/kubelet/pods/b4bae750-8bf9-4804-bab7-9c94d4be7bb6/volumes" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.059885 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch"] Feb 20 08:27:03 crc kubenswrapper[4948]: E0220 08:27:03.060920 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1847849-cdf4-4ea6-a540-c01503a003b2" containerName="init" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.060935 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1847849-cdf4-4ea6-a540-c01503a003b2" containerName="init" Feb 20 08:27:03 crc kubenswrapper[4948]: E0220 08:27:03.060951 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" containerName="dnsmasq-dns" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.060958 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" containerName="dnsmasq-dns" Feb 20 08:27:03 crc kubenswrapper[4948]: E0220 08:27:03.060996 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" containerName="init" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.061006 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" containerName="init" Feb 20 08:27:03 crc kubenswrapper[4948]: E0220 08:27:03.061023 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1847849-cdf4-4ea6-a540-c01503a003b2" containerName="dnsmasq-dns" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.061030 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1847849-cdf4-4ea6-a540-c01503a003b2" containerName="dnsmasq-dns" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.061241 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4bae750-8bf9-4804-bab7-9c94d4be7bb6" containerName="dnsmasq-dns" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.061274 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1847849-cdf4-4ea6-a540-c01503a003b2" containerName="dnsmasq-dns" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.061942 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.065179 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.066921 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.069586 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.070244 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.094095 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch"] Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.160037 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.160225 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.160302 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.160328 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2krxb\" (UniqueName: \"kubernetes.io/projected/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-kube-api-access-2krxb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.262339 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.262395 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2krxb\" (UniqueName: \"kubernetes.io/projected/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-kube-api-access-2krxb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.262563 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.262669 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.268887 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.269590 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.271802 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.281562 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2krxb\" (UniqueName: \"kubernetes.io/projected/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-kube-api-access-2krxb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.409593 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.896578 4948 generic.go:334] "Generic (PLEG): container finished" podID="037b329f-712f-4a67-984a-75affd2a57b7" containerID="9fcd3d9744d41147b3cfe90374517e1e56f49253e12cf9785df1944e69f8e3a2" exitCode=0 Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.896639 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"037b329f-712f-4a67-984a-75affd2a57b7","Type":"ContainerDied","Data":"9fcd3d9744d41147b3cfe90374517e1e56f49253e12cf9785df1944e69f8e3a2"} Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.903642 4948 generic.go:334] "Generic (PLEG): container finished" podID="f063bb62-10c3-46d9-a41d-56d2826059e0" containerID="878f9968270125c43cab5cdbc68ca06efa6e152bfb9c605c96475afc50f8ad38" exitCode=0 Feb 20 08:27:03 crc kubenswrapper[4948]: I0220 08:27:03.903689 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f063bb62-10c3-46d9-a41d-56d2826059e0","Type":"ContainerDied","Data":"878f9968270125c43cab5cdbc68ca06efa6e152bfb9c605c96475afc50f8ad38"} Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.025074 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch"] Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.025764 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.915849 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"037b329f-712f-4a67-984a-75affd2a57b7","Type":"ContainerStarted","Data":"9f638c6325085bb3f636e25b3784184a35ec0ac8c3fe43494b719b19a5a49e3a"} Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.916413 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.917714 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" event={"ID":"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f","Type":"ContainerStarted","Data":"e390522c20f722c1095e8ce15bfbab5e5bebdf33d9f0c27bcc6e6d315e7becc5"} Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.930006 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f063bb62-10c3-46d9-a41d-56d2826059e0","Type":"ContainerStarted","Data":"0366bfe9eabe6a009ea2dc265496f54894233021b297a92a54a5dc68da912163"} Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.930381 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.940272 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.94025592 podStartE2EDuration="36.94025592s" podCreationTimestamp="2026-02-20 08:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:27:04.936865026 +0000 UTC m=+1273.911359846" watchObservedRunningTime="2026-02-20 08:27:04.94025592 +0000 UTC m=+1273.914750740" Feb 20 08:27:04 crc kubenswrapper[4948]: I0220 08:27:04.969255 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.969236705 podStartE2EDuration="37.969236705s" podCreationTimestamp="2026-02-20 08:26:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:27:04.966820825 +0000 UTC m=+1273.941315645" watchObservedRunningTime="2026-02-20 08:27:04.969236705 +0000 UTC m=+1273.943731525" Feb 20 08:27:08 crc kubenswrapper[4948]: I0220 08:27:08.025012 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:27:08 crc kubenswrapper[4948]: I0220 08:27:08.025376 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:27:14 crc kubenswrapper[4948]: I0220 08:27:14.028693 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" event={"ID":"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f","Type":"ContainerStarted","Data":"db62d9363c40e171b052da3433783d0ee6d1a1d0ff212dae6ec69ed654f36fb5"} Feb 20 08:27:18 crc kubenswrapper[4948]: I0220 08:27:18.132304 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 20 08:27:18 crc kubenswrapper[4948]: I0220 08:27:18.162518 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" podStartSLOduration=5.708976495 podStartE2EDuration="15.162500021s" podCreationTimestamp="2026-02-20 08:27:03 +0000 UTC" firstStartedPulling="2026-02-20 08:27:04.025499229 +0000 UTC m=+1272.999994059" lastFinishedPulling="2026-02-20 08:27:13.479022735 +0000 UTC m=+1282.453517585" observedRunningTime="2026-02-20 08:27:14.060010871 +0000 UTC m=+1283.034505731" watchObservedRunningTime="2026-02-20 08:27:18.162500021 +0000 UTC m=+1287.136994831" Feb 20 08:27:18 crc kubenswrapper[4948]: I0220 08:27:18.825134 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 20 08:27:24 crc kubenswrapper[4948]: I0220 08:27:24.186269 4948 generic.go:334] "Generic (PLEG): container finished" podID="90b6e7df-b3e1-4d05-a6b3-7fe47d70804f" containerID="db62d9363c40e171b052da3433783d0ee6d1a1d0ff212dae6ec69ed654f36fb5" exitCode=0 Feb 20 08:27:24 crc kubenswrapper[4948]: I0220 08:27:24.186368 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" event={"ID":"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f","Type":"ContainerDied","Data":"db62d9363c40e171b052da3433783d0ee6d1a1d0ff212dae6ec69ed654f36fb5"} Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.674370 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.729481 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2krxb\" (UniqueName: \"kubernetes.io/projected/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-kube-api-access-2krxb\") pod \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.729680 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-repo-setup-combined-ca-bundle\") pod \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.729708 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-inventory\") pod \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.729750 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-ssh-key-openstack-edpm-ipam\") pod \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\" (UID: \"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f\") " Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.735778 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "90b6e7df-b3e1-4d05-a6b3-7fe47d70804f" (UID: "90b6e7df-b3e1-4d05-a6b3-7fe47d70804f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.740168 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-kube-api-access-2krxb" (OuterVolumeSpecName: "kube-api-access-2krxb") pod "90b6e7df-b3e1-4d05-a6b3-7fe47d70804f" (UID: "90b6e7df-b3e1-4d05-a6b3-7fe47d70804f"). InnerVolumeSpecName "kube-api-access-2krxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.762325 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "90b6e7df-b3e1-4d05-a6b3-7fe47d70804f" (UID: "90b6e7df-b3e1-4d05-a6b3-7fe47d70804f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.769085 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-inventory" (OuterVolumeSpecName: "inventory") pod "90b6e7df-b3e1-4d05-a6b3-7fe47d70804f" (UID: "90b6e7df-b3e1-4d05-a6b3-7fe47d70804f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.833686 4948 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.833742 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.833758 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:27:25 crc kubenswrapper[4948]: I0220 08:27:25.833773 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2krxb\" (UniqueName: \"kubernetes.io/projected/90b6e7df-b3e1-4d05-a6b3-7fe47d70804f-kube-api-access-2krxb\") on node \"crc\" DevicePath \"\"" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.211257 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" event={"ID":"90b6e7df-b3e1-4d05-a6b3-7fe47d70804f","Type":"ContainerDied","Data":"e390522c20f722c1095e8ce15bfbab5e5bebdf33d9f0c27bcc6e6d315e7becc5"} Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.211617 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e390522c20f722c1095e8ce15bfbab5e5bebdf33d9f0c27bcc6e6d315e7becc5" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.211349 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.335418 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9"] Feb 20 08:27:26 crc kubenswrapper[4948]: E0220 08:27:26.335893 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90b6e7df-b3e1-4d05-a6b3-7fe47d70804f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.335914 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="90b6e7df-b3e1-4d05-a6b3-7fe47d70804f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.336137 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="90b6e7df-b3e1-4d05-a6b3-7fe47d70804f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.337036 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.339348 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.339865 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.340074 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.344595 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.349141 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9"] Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.450338 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdjs4\" (UniqueName: \"kubernetes.io/projected/c3e7cce4-5144-4190-8013-83428c66cde9-kube-api-access-sdjs4\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.450693 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.451056 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.553058 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.553473 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdjs4\" (UniqueName: \"kubernetes.io/projected/c3e7cce4-5144-4190-8013-83428c66cde9-kube-api-access-sdjs4\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.553617 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.561423 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.562106 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.570770 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdjs4\" (UniqueName: \"kubernetes.io/projected/c3e7cce4-5144-4190-8013-83428c66cde9-kube-api-access-sdjs4\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-f7ww9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:26 crc kubenswrapper[4948]: I0220 08:27:26.669739 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:27 crc kubenswrapper[4948]: I0220 08:27:27.290419 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9"] Feb 20 08:27:28 crc kubenswrapper[4948]: I0220 08:27:28.240919 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" event={"ID":"c3e7cce4-5144-4190-8013-83428c66cde9","Type":"ContainerStarted","Data":"3c99aa66568f6e79e574e82f764277e758470002a35a2b965f15c2be5d2ddbe6"} Feb 20 08:27:28 crc kubenswrapper[4948]: I0220 08:27:28.241409 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" event={"ID":"c3e7cce4-5144-4190-8013-83428c66cde9","Type":"ContainerStarted","Data":"92919d350e3d8fc6a7b429f20d4d8fc554e212055abe58d5ab447af9034b6250"} Feb 20 08:27:28 crc kubenswrapper[4948]: I0220 08:27:28.271717 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" podStartSLOduration=1.8622712940000001 podStartE2EDuration="2.271693857s" podCreationTimestamp="2026-02-20 08:27:26 +0000 UTC" firstStartedPulling="2026-02-20 08:27:27.299451976 +0000 UTC m=+1296.273946796" lastFinishedPulling="2026-02-20 08:27:27.708874539 +0000 UTC m=+1296.683369359" observedRunningTime="2026-02-20 08:27:28.264502699 +0000 UTC m=+1297.238997529" watchObservedRunningTime="2026-02-20 08:27:28.271693857 +0000 UTC m=+1297.246188687" Feb 20 08:27:31 crc kubenswrapper[4948]: I0220 08:27:31.290323 4948 generic.go:334] "Generic (PLEG): container finished" podID="c3e7cce4-5144-4190-8013-83428c66cde9" containerID="3c99aa66568f6e79e574e82f764277e758470002a35a2b965f15c2be5d2ddbe6" exitCode=0 Feb 20 08:27:31 crc kubenswrapper[4948]: I0220 08:27:31.290544 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" event={"ID":"c3e7cce4-5144-4190-8013-83428c66cde9","Type":"ContainerDied","Data":"3c99aa66568f6e79e574e82f764277e758470002a35a2b965f15c2be5d2ddbe6"} Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.779582 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.891618 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-ssh-key-openstack-edpm-ipam\") pod \"c3e7cce4-5144-4190-8013-83428c66cde9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.892130 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-inventory\") pod \"c3e7cce4-5144-4190-8013-83428c66cde9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.892201 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdjs4\" (UniqueName: \"kubernetes.io/projected/c3e7cce4-5144-4190-8013-83428c66cde9-kube-api-access-sdjs4\") pod \"c3e7cce4-5144-4190-8013-83428c66cde9\" (UID: \"c3e7cce4-5144-4190-8013-83428c66cde9\") " Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.901570 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3e7cce4-5144-4190-8013-83428c66cde9-kube-api-access-sdjs4" (OuterVolumeSpecName: "kube-api-access-sdjs4") pod "c3e7cce4-5144-4190-8013-83428c66cde9" (UID: "c3e7cce4-5144-4190-8013-83428c66cde9"). InnerVolumeSpecName "kube-api-access-sdjs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.937045 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c3e7cce4-5144-4190-8013-83428c66cde9" (UID: "c3e7cce4-5144-4190-8013-83428c66cde9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.944967 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-inventory" (OuterVolumeSpecName: "inventory") pod "c3e7cce4-5144-4190-8013-83428c66cde9" (UID: "c3e7cce4-5144-4190-8013-83428c66cde9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.996035 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.996097 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdjs4\" (UniqueName: \"kubernetes.io/projected/c3e7cce4-5144-4190-8013-83428c66cde9-kube-api-access-sdjs4\") on node \"crc\" DevicePath \"\"" Feb 20 08:27:32 crc kubenswrapper[4948]: I0220 08:27:32.996121 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c3e7cce4-5144-4190-8013-83428c66cde9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.317787 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" event={"ID":"c3e7cce4-5144-4190-8013-83428c66cde9","Type":"ContainerDied","Data":"92919d350e3d8fc6a7b429f20d4d8fc554e212055abe58d5ab447af9034b6250"} Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.318227 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92919d350e3d8fc6a7b429f20d4d8fc554e212055abe58d5ab447af9034b6250" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.317859 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-f7ww9" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.432225 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g"] Feb 20 08:27:33 crc kubenswrapper[4948]: E0220 08:27:33.433230 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3e7cce4-5144-4190-8013-83428c66cde9" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.433261 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3e7cce4-5144-4190-8013-83428c66cde9" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.433597 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3e7cce4-5144-4190-8013-83428c66cde9" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.434689 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.441678 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.441869 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.441950 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.442049 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.448177 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g"] Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.508191 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.508291 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.508491 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljn9d\" (UniqueName: \"kubernetes.io/projected/e9bed94f-dc85-433b-b2db-8da400959f54-kube-api-access-ljn9d\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.508865 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.611155 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.611270 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljn9d\" (UniqueName: \"kubernetes.io/projected/e9bed94f-dc85-433b-b2db-8da400959f54-kube-api-access-ljn9d\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.611414 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.611554 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.616812 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.623451 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.623932 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.634935 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljn9d\" (UniqueName: \"kubernetes.io/projected/e9bed94f-dc85-433b-b2db-8da400959f54-kube-api-access-ljn9d\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4265g\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:33 crc kubenswrapper[4948]: I0220 08:27:33.767446 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:27:34 crc kubenswrapper[4948]: I0220 08:27:34.342237 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g"] Feb 20 08:27:34 crc kubenswrapper[4948]: W0220 08:27:34.347437 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9bed94f_dc85_433b_b2db_8da400959f54.slice/crio-569d95f896c3d860d075c4d7112372cdd51f62d88d362789225ed9ac1821cb97 WatchSource:0}: Error finding container 569d95f896c3d860d075c4d7112372cdd51f62d88d362789225ed9ac1821cb97: Status 404 returned error can't find the container with id 569d95f896c3d860d075c4d7112372cdd51f62d88d362789225ed9ac1821cb97 Feb 20 08:27:35 crc kubenswrapper[4948]: I0220 08:27:35.351665 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" event={"ID":"e9bed94f-dc85-433b-b2db-8da400959f54","Type":"ContainerStarted","Data":"966bce30bb61203264725c1e50c5639d590639efbe4b4a138802094aa4f1d142"} Feb 20 08:27:35 crc kubenswrapper[4948]: I0220 08:27:35.351944 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" event={"ID":"e9bed94f-dc85-433b-b2db-8da400959f54","Type":"ContainerStarted","Data":"569d95f896c3d860d075c4d7112372cdd51f62d88d362789225ed9ac1821cb97"} Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.025063 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.025778 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.025846 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.026727 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d9efd6716995fcdcebdcf01e76fc7927735e34b1da4918444f90e59bd0f6d815"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.026823 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://d9efd6716995fcdcebdcf01e76fc7927735e34b1da4918444f90e59bd0f6d815" gracePeriod=600 Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.390667 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="d9efd6716995fcdcebdcf01e76fc7927735e34b1da4918444f90e59bd0f6d815" exitCode=0 Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.390758 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"d9efd6716995fcdcebdcf01e76fc7927735e34b1da4918444f90e59bd0f6d815"} Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.391482 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43"} Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.391546 4948 scope.go:117] "RemoveContainer" containerID="8f0b8a806330c630468951158113ee0ecaad5dbddb81b09c063752a17f21282c" Feb 20 08:27:38 crc kubenswrapper[4948]: I0220 08:27:38.424231 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" podStartSLOduration=5.00794339 podStartE2EDuration="5.424211271s" podCreationTimestamp="2026-02-20 08:27:33 +0000 UTC" firstStartedPulling="2026-02-20 08:27:34.350785309 +0000 UTC m=+1303.325280129" lastFinishedPulling="2026-02-20 08:27:34.76705318 +0000 UTC m=+1303.741548010" observedRunningTime="2026-02-20 08:27:35.376701823 +0000 UTC m=+1304.351196703" watchObservedRunningTime="2026-02-20 08:27:38.424211271 +0000 UTC m=+1307.398706091" Feb 20 08:27:57 crc kubenswrapper[4948]: I0220 08:27:57.961889 4948 scope.go:117] "RemoveContainer" containerID="562bf68215847fe83f99073d009b1ee89904610f3d702b2ab282baecbca78fb2" Feb 20 08:27:58 crc kubenswrapper[4948]: I0220 08:27:58.002130 4948 scope.go:117] "RemoveContainer" containerID="d4c99094574230bbe4d4f5bd6cf1019d664fb9f16355c0c0e8d85cd644e73d32" Feb 20 08:27:58 crc kubenswrapper[4948]: I0220 08:27:58.047844 4948 scope.go:117] "RemoveContainer" containerID="48592a00a61a9ce774a176445c59d2e205743cbfe0a6d5b932c49a8ff31d312a" Feb 20 08:27:58 crc kubenswrapper[4948]: I0220 08:27:58.084311 4948 scope.go:117] "RemoveContainer" containerID="c4af4c0a8310dde22c0962922d9e1a2de48ad4ce8cecfbf27d7c0a78b93a2059" Feb 20 08:27:58 crc kubenswrapper[4948]: I0220 08:27:58.125330 4948 scope.go:117] "RemoveContainer" containerID="5dfd385e001fae733c595da928ba9b46f6e3cc2e4d358e37840e650d21ce855a" Feb 20 08:28:58 crc kubenswrapper[4948]: I0220 08:28:58.274456 4948 scope.go:117] "RemoveContainer" containerID="656e84496ca041e8a454b07ed67e165b130af1030016564ae4519b8e2e6f11e5" Feb 20 08:28:58 crc kubenswrapper[4948]: I0220 08:28:58.332901 4948 scope.go:117] "RemoveContainer" containerID="577865ca69f0d8457d617948e5cac2be0a124e4620edc60df1395130de8a2334" Feb 20 08:28:58 crc kubenswrapper[4948]: I0220 08:28:58.355680 4948 scope.go:117] "RemoveContainer" containerID="a1b60a312d88b84db5225c1aff5290457499b2775b6fffec7ca3246896b28523" Feb 20 08:28:58 crc kubenswrapper[4948]: I0220 08:28:58.593685 4948 scope.go:117] "RemoveContainer" containerID="fa275de1b67d8651c437fb815c4f8470dc8d77cd9885aaddee1a7419f33fd25c" Feb 20 08:28:58 crc kubenswrapper[4948]: I0220 08:28:58.656936 4948 scope.go:117] "RemoveContainer" containerID="344f41da66c91718be93d872d218e57c352e04b55c97d3dc69d71b942dbe58d1" Feb 20 08:28:58 crc kubenswrapper[4948]: I0220 08:28:58.682561 4948 scope.go:117] "RemoveContainer" containerID="a66ddb00a7f57dc492a3a8a03a3bce06f45c65381b1e5db4f923e10fec5267a0" Feb 20 08:28:58 crc kubenswrapper[4948]: I0220 08:28:58.714102 4948 scope.go:117] "RemoveContainer" containerID="e38f809977ba5076c8af037b7e38d52d33545f893ce90480e346caeaaef81dee" Feb 20 08:29:38 crc kubenswrapper[4948]: I0220 08:29:38.025596 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:29:38 crc kubenswrapper[4948]: I0220 08:29:38.027238 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.151947 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm"] Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.153870 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.155720 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.165044 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.178993 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm"] Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.310425 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpkkr\" (UniqueName: \"kubernetes.io/projected/2aaa3d0b-4371-404a-8d94-9a7158fff416-kube-api-access-hpkkr\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.310516 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2aaa3d0b-4371-404a-8d94-9a7158fff416-secret-volume\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.310597 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2aaa3d0b-4371-404a-8d94-9a7158fff416-config-volume\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.412799 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2aaa3d0b-4371-404a-8d94-9a7158fff416-config-volume\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.413112 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpkkr\" (UniqueName: \"kubernetes.io/projected/2aaa3d0b-4371-404a-8d94-9a7158fff416-kube-api-access-hpkkr\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.413156 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2aaa3d0b-4371-404a-8d94-9a7158fff416-secret-volume\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.414374 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2aaa3d0b-4371-404a-8d94-9a7158fff416-config-volume\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.419120 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2aaa3d0b-4371-404a-8d94-9a7158fff416-secret-volume\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.433017 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpkkr\" (UniqueName: \"kubernetes.io/projected/2aaa3d0b-4371-404a-8d94-9a7158fff416-kube-api-access-hpkkr\") pod \"collect-profiles-29526270-98wxm\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.473949 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:00 crc kubenswrapper[4948]: I0220 08:30:00.986612 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm"] Feb 20 08:30:01 crc kubenswrapper[4948]: I0220 08:30:01.138683 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" event={"ID":"2aaa3d0b-4371-404a-8d94-9a7158fff416","Type":"ContainerStarted","Data":"34ee96251b37f911907f4a06d490080e0cb865adf32db4a0b0cd8c0d113ad82a"} Feb 20 08:30:02 crc kubenswrapper[4948]: I0220 08:30:02.153592 4948 generic.go:334] "Generic (PLEG): container finished" podID="2aaa3d0b-4371-404a-8d94-9a7158fff416" containerID="21765550ece8828e59342442ce6d5be7d08bd1dbcd30f8abbcc286c1bfd5f713" exitCode=0 Feb 20 08:30:02 crc kubenswrapper[4948]: I0220 08:30:02.153722 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" event={"ID":"2aaa3d0b-4371-404a-8d94-9a7158fff416","Type":"ContainerDied","Data":"21765550ece8828e59342442ce6d5be7d08bd1dbcd30f8abbcc286c1bfd5f713"} Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.503152 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.675943 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpkkr\" (UniqueName: \"kubernetes.io/projected/2aaa3d0b-4371-404a-8d94-9a7158fff416-kube-api-access-hpkkr\") pod \"2aaa3d0b-4371-404a-8d94-9a7158fff416\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.676086 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2aaa3d0b-4371-404a-8d94-9a7158fff416-secret-volume\") pod \"2aaa3d0b-4371-404a-8d94-9a7158fff416\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.676214 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2aaa3d0b-4371-404a-8d94-9a7158fff416-config-volume\") pod \"2aaa3d0b-4371-404a-8d94-9a7158fff416\" (UID: \"2aaa3d0b-4371-404a-8d94-9a7158fff416\") " Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.677122 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2aaa3d0b-4371-404a-8d94-9a7158fff416-config-volume" (OuterVolumeSpecName: "config-volume") pod "2aaa3d0b-4371-404a-8d94-9a7158fff416" (UID: "2aaa3d0b-4371-404a-8d94-9a7158fff416"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.685890 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aaa3d0b-4371-404a-8d94-9a7158fff416-kube-api-access-hpkkr" (OuterVolumeSpecName: "kube-api-access-hpkkr") pod "2aaa3d0b-4371-404a-8d94-9a7158fff416" (UID: "2aaa3d0b-4371-404a-8d94-9a7158fff416"). InnerVolumeSpecName "kube-api-access-hpkkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.686127 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aaa3d0b-4371-404a-8d94-9a7158fff416-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2aaa3d0b-4371-404a-8d94-9a7158fff416" (UID: "2aaa3d0b-4371-404a-8d94-9a7158fff416"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.778825 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpkkr\" (UniqueName: \"kubernetes.io/projected/2aaa3d0b-4371-404a-8d94-9a7158fff416-kube-api-access-hpkkr\") on node \"crc\" DevicePath \"\"" Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.778877 4948 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2aaa3d0b-4371-404a-8d94-9a7158fff416-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:30:03 crc kubenswrapper[4948]: I0220 08:30:03.778898 4948 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2aaa3d0b-4371-404a-8d94-9a7158fff416-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:30:04 crc kubenswrapper[4948]: I0220 08:30:04.181960 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" event={"ID":"2aaa3d0b-4371-404a-8d94-9a7158fff416","Type":"ContainerDied","Data":"34ee96251b37f911907f4a06d490080e0cb865adf32db4a0b0cd8c0d113ad82a"} Feb 20 08:30:04 crc kubenswrapper[4948]: I0220 08:30:04.182215 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34ee96251b37f911907f4a06d490080e0cb865adf32db4a0b0cd8c0d113ad82a" Feb 20 08:30:04 crc kubenswrapper[4948]: I0220 08:30:04.182560 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm" Feb 20 08:30:08 crc kubenswrapper[4948]: I0220 08:30:08.024746 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:30:08 crc kubenswrapper[4948]: I0220 08:30:08.025424 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:30:26 crc kubenswrapper[4948]: I0220 08:30:26.404256 4948 generic.go:334] "Generic (PLEG): container finished" podID="e9bed94f-dc85-433b-b2db-8da400959f54" containerID="966bce30bb61203264725c1e50c5639d590639efbe4b4a138802094aa4f1d142" exitCode=0 Feb 20 08:30:26 crc kubenswrapper[4948]: I0220 08:30:26.404372 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" event={"ID":"e9bed94f-dc85-433b-b2db-8da400959f54","Type":"ContainerDied","Data":"966bce30bb61203264725c1e50c5639d590639efbe4b4a138802094aa4f1d142"} Feb 20 08:30:27 crc kubenswrapper[4948]: I0220 08:30:27.913182 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.086391 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-inventory\") pod \"e9bed94f-dc85-433b-b2db-8da400959f54\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.086465 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-ssh-key-openstack-edpm-ipam\") pod \"e9bed94f-dc85-433b-b2db-8da400959f54\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.086547 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-bootstrap-combined-ca-bundle\") pod \"e9bed94f-dc85-433b-b2db-8da400959f54\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.086577 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljn9d\" (UniqueName: \"kubernetes.io/projected/e9bed94f-dc85-433b-b2db-8da400959f54-kube-api-access-ljn9d\") pod \"e9bed94f-dc85-433b-b2db-8da400959f54\" (UID: \"e9bed94f-dc85-433b-b2db-8da400959f54\") " Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.092169 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "e9bed94f-dc85-433b-b2db-8da400959f54" (UID: "e9bed94f-dc85-433b-b2db-8da400959f54"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.097966 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9bed94f-dc85-433b-b2db-8da400959f54-kube-api-access-ljn9d" (OuterVolumeSpecName: "kube-api-access-ljn9d") pod "e9bed94f-dc85-433b-b2db-8da400959f54" (UID: "e9bed94f-dc85-433b-b2db-8da400959f54"). InnerVolumeSpecName "kube-api-access-ljn9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.113590 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-inventory" (OuterVolumeSpecName: "inventory") pod "e9bed94f-dc85-433b-b2db-8da400959f54" (UID: "e9bed94f-dc85-433b-b2db-8da400959f54"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.117552 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e9bed94f-dc85-433b-b2db-8da400959f54" (UID: "e9bed94f-dc85-433b-b2db-8da400959f54"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.188504 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.188549 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.188560 4948 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bed94f-dc85-433b-b2db-8da400959f54-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.188569 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljn9d\" (UniqueName: \"kubernetes.io/projected/e9bed94f-dc85-433b-b2db-8da400959f54-kube-api-access-ljn9d\") on node \"crc\" DevicePath \"\"" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.436641 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" event={"ID":"e9bed94f-dc85-433b-b2db-8da400959f54","Type":"ContainerDied","Data":"569d95f896c3d860d075c4d7112372cdd51f62d88d362789225ed9ac1821cb97"} Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.436935 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="569d95f896c3d860d075c4d7112372cdd51f62d88d362789225ed9ac1821cb97" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.436718 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4265g" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.542147 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll"] Feb 20 08:30:28 crc kubenswrapper[4948]: E0220 08:30:28.542558 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aaa3d0b-4371-404a-8d94-9a7158fff416" containerName="collect-profiles" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.542574 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aaa3d0b-4371-404a-8d94-9a7158fff416" containerName="collect-profiles" Feb 20 08:30:28 crc kubenswrapper[4948]: E0220 08:30:28.542614 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9bed94f-dc85-433b-b2db-8da400959f54" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.542623 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9bed94f-dc85-433b-b2db-8da400959f54" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.542796 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9bed94f-dc85-433b-b2db-8da400959f54" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.542816 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aaa3d0b-4371-404a-8d94-9a7158fff416" containerName="collect-profiles" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.543707 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.547949 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.548473 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.548714 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.548920 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.567651 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll"] Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.697845 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgbgc\" (UniqueName: \"kubernetes.io/projected/5864a68d-650e-4bcf-b705-619c0f27445b-kube-api-access-lgbgc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.697965 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.698167 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.800434 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.800555 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.800831 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgbgc\" (UniqueName: \"kubernetes.io/projected/5864a68d-650e-4bcf-b705-619c0f27445b-kube-api-access-lgbgc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.807219 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.808759 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.831187 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgbgc\" (UniqueName: \"kubernetes.io/projected/5864a68d-650e-4bcf-b705-619c0f27445b-kube-api-access-lgbgc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-mhkll\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:28 crc kubenswrapper[4948]: I0220 08:30:28.872905 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:30:29 crc kubenswrapper[4948]: I0220 08:30:29.473436 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll"] Feb 20 08:30:29 crc kubenswrapper[4948]: W0220 08:30:29.487390 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5864a68d_650e_4bcf_b705_619c0f27445b.slice/crio-3ff23193cbb1aafb8b5d373010ecf579e6350e550af86a1ff2fe999d6f3221b3 WatchSource:0}: Error finding container 3ff23193cbb1aafb8b5d373010ecf579e6350e550af86a1ff2fe999d6f3221b3: Status 404 returned error can't find the container with id 3ff23193cbb1aafb8b5d373010ecf579e6350e550af86a1ff2fe999d6f3221b3 Feb 20 08:30:30 crc kubenswrapper[4948]: I0220 08:30:30.457290 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" event={"ID":"5864a68d-650e-4bcf-b705-619c0f27445b","Type":"ContainerStarted","Data":"bb15461fc1a5607b927c4774ccf8aa52aa2b1f66a0749cf932076cda97280dfb"} Feb 20 08:30:30 crc kubenswrapper[4948]: I0220 08:30:30.457863 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" event={"ID":"5864a68d-650e-4bcf-b705-619c0f27445b","Type":"ContainerStarted","Data":"3ff23193cbb1aafb8b5d373010ecf579e6350e550af86a1ff2fe999d6f3221b3"} Feb 20 08:30:30 crc kubenswrapper[4948]: I0220 08:30:30.480546 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" podStartSLOduration=1.968606227 podStartE2EDuration="2.480524084s" podCreationTimestamp="2026-02-20 08:30:28 +0000 UTC" firstStartedPulling="2026-02-20 08:30:29.490123109 +0000 UTC m=+1478.464617929" lastFinishedPulling="2026-02-20 08:30:30.002040966 +0000 UTC m=+1478.976535786" observedRunningTime="2026-02-20 08:30:30.470715932 +0000 UTC m=+1479.445210752" watchObservedRunningTime="2026-02-20 08:30:30.480524084 +0000 UTC m=+1479.455018904" Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.025710 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.026557 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.026681 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.028033 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.028403 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" gracePeriod=600 Feb 20 08:30:38 crc kubenswrapper[4948]: E0220 08:30:38.154015 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.540311 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" exitCode=0 Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.540371 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43"} Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.540417 4948 scope.go:117] "RemoveContainer" containerID="d9efd6716995fcdcebdcf01e76fc7927735e34b1da4918444f90e59bd0f6d815" Feb 20 08:30:38 crc kubenswrapper[4948]: I0220 08:30:38.541267 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:30:38 crc kubenswrapper[4948]: E0220 08:30:38.541724 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:30:52 crc kubenswrapper[4948]: I0220 08:30:52.723197 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:30:52 crc kubenswrapper[4948]: E0220 08:30:52.725563 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:31:06 crc kubenswrapper[4948]: I0220 08:31:06.722951 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:31:06 crc kubenswrapper[4948]: E0220 08:31:06.723915 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:31:21 crc kubenswrapper[4948]: I0220 08:31:21.730231 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:31:21 crc kubenswrapper[4948]: E0220 08:31:21.731040 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:31:34 crc kubenswrapper[4948]: I0220 08:31:34.883503 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bdphs"] Feb 20 08:31:34 crc kubenswrapper[4948]: I0220 08:31:34.886606 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:34 crc kubenswrapper[4948]: I0220 08:31:34.901490 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bdphs"] Feb 20 08:31:34 crc kubenswrapper[4948]: I0220 08:31:34.998346 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-catalog-content\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:34 crc kubenswrapper[4948]: I0220 08:31:34.998414 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l4qt\" (UniqueName: \"kubernetes.io/projected/060989bf-ef06-4060-bc25-915180f42e5b-kube-api-access-4l4qt\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:34 crc kubenswrapper[4948]: I0220 08:31:34.998494 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-utilities\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:35 crc kubenswrapper[4948]: I0220 08:31:35.100545 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-catalog-content\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:35 crc kubenswrapper[4948]: I0220 08:31:35.100697 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l4qt\" (UniqueName: \"kubernetes.io/projected/060989bf-ef06-4060-bc25-915180f42e5b-kube-api-access-4l4qt\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:35 crc kubenswrapper[4948]: I0220 08:31:35.100933 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-utilities\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:35 crc kubenswrapper[4948]: I0220 08:31:35.101065 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-catalog-content\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:35 crc kubenswrapper[4948]: I0220 08:31:35.101584 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-utilities\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:35 crc kubenswrapper[4948]: I0220 08:31:35.120515 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l4qt\" (UniqueName: \"kubernetes.io/projected/060989bf-ef06-4060-bc25-915180f42e5b-kube-api-access-4l4qt\") pod \"redhat-operators-bdphs\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:35 crc kubenswrapper[4948]: I0220 08:31:35.256669 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:35 crc kubenswrapper[4948]: I0220 08:31:35.715275 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bdphs"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.050405 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-bv65d"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.055779 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-eb36-account-create-update-554g5"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.064749 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6d58-account-create-update-bbx44"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.076611 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-gm6g6"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.087538 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-bv65d"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.095016 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-eb36-account-create-update-554g5"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.120633 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-6d58-account-create-update-bbx44"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.138198 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-gm6g6"] Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.145676 4948 generic.go:334] "Generic (PLEG): container finished" podID="060989bf-ef06-4060-bc25-915180f42e5b" containerID="cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1" exitCode=0 Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.145738 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdphs" event={"ID":"060989bf-ef06-4060-bc25-915180f42e5b","Type":"ContainerDied","Data":"cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1"} Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.145764 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdphs" event={"ID":"060989bf-ef06-4060-bc25-915180f42e5b","Type":"ContainerStarted","Data":"53735cceaab8db6cbf509cf3d7a338fd07617ca1b94b6981218603729cf70256"} Feb 20 08:31:36 crc kubenswrapper[4948]: I0220 08:31:36.723539 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:31:36 crc kubenswrapper[4948]: E0220 08:31:36.723945 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:31:37 crc kubenswrapper[4948]: I0220 08:31:37.156534 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdphs" event={"ID":"060989bf-ef06-4060-bc25-915180f42e5b","Type":"ContainerStarted","Data":"7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966"} Feb 20 08:31:37 crc kubenswrapper[4948]: I0220 08:31:37.736221 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52238644-5964-427e-8b77-9ca228d031db" path="/var/lib/kubelet/pods/52238644-5964-427e-8b77-9ca228d031db/volumes" Feb 20 08:31:37 crc kubenswrapper[4948]: I0220 08:31:37.738563 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5bf7a21-af5d-4334-bc98-bd0f9eb9035a" path="/var/lib/kubelet/pods/b5bf7a21-af5d-4334-bc98-bd0f9eb9035a/volumes" Feb 20 08:31:37 crc kubenswrapper[4948]: I0220 08:31:37.739957 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb" path="/var/lib/kubelet/pods/e1700d74-b484-4c6e-8fd3-d9dab8f4b7eb/volumes" Feb 20 08:31:37 crc kubenswrapper[4948]: I0220 08:31:37.741223 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f917fa7f-51f9-4c8f-bee2-11529a6d44a6" path="/var/lib/kubelet/pods/f917fa7f-51f9-4c8f-bee2-11529a6d44a6/volumes" Feb 20 08:31:38 crc kubenswrapper[4948]: I0220 08:31:38.040791 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-fs9rr"] Feb 20 08:31:38 crc kubenswrapper[4948]: I0220 08:31:38.055305 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-c08d-account-create-update-54vb7"] Feb 20 08:31:38 crc kubenswrapper[4948]: I0220 08:31:38.067434 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-fs9rr"] Feb 20 08:31:38 crc kubenswrapper[4948]: I0220 08:31:38.084195 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-c08d-account-create-update-54vb7"] Feb 20 08:31:38 crc kubenswrapper[4948]: I0220 08:31:38.167879 4948 generic.go:334] "Generic (PLEG): container finished" podID="060989bf-ef06-4060-bc25-915180f42e5b" containerID="7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966" exitCode=0 Feb 20 08:31:38 crc kubenswrapper[4948]: I0220 08:31:38.167927 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdphs" event={"ID":"060989bf-ef06-4060-bc25-915180f42e5b","Type":"ContainerDied","Data":"7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966"} Feb 20 08:31:39 crc kubenswrapper[4948]: I0220 08:31:39.179496 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdphs" event={"ID":"060989bf-ef06-4060-bc25-915180f42e5b","Type":"ContainerStarted","Data":"37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529"} Feb 20 08:31:39 crc kubenswrapper[4948]: I0220 08:31:39.205194 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bdphs" podStartSLOduration=2.787117993 podStartE2EDuration="5.205172714s" podCreationTimestamp="2026-02-20 08:31:34 +0000 UTC" firstStartedPulling="2026-02-20 08:31:36.152658963 +0000 UTC m=+1545.127153783" lastFinishedPulling="2026-02-20 08:31:38.570713684 +0000 UTC m=+1547.545208504" observedRunningTime="2026-02-20 08:31:39.195878464 +0000 UTC m=+1548.170373284" watchObservedRunningTime="2026-02-20 08:31:39.205172714 +0000 UTC m=+1548.179667544" Feb 20 08:31:39 crc kubenswrapper[4948]: I0220 08:31:39.740178 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="333ba21a-82a9-4f39-9ec5-afb7ae2ccf60" path="/var/lib/kubelet/pods/333ba21a-82a9-4f39-9ec5-afb7ae2ccf60/volumes" Feb 20 08:31:39 crc kubenswrapper[4948]: I0220 08:31:39.741584 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d815209-521c-4c7f-a026-18899832459f" path="/var/lib/kubelet/pods/4d815209-521c-4c7f-a026-18899832459f/volumes" Feb 20 08:31:45 crc kubenswrapper[4948]: I0220 08:31:45.257506 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:45 crc kubenswrapper[4948]: I0220 08:31:45.258235 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:46 crc kubenswrapper[4948]: I0220 08:31:46.328623 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bdphs" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="registry-server" probeResult="failure" output=< Feb 20 08:31:46 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 08:31:46 crc kubenswrapper[4948]: > Feb 20 08:31:49 crc kubenswrapper[4948]: I0220 08:31:49.296733 4948 generic.go:334] "Generic (PLEG): container finished" podID="5864a68d-650e-4bcf-b705-619c0f27445b" containerID="bb15461fc1a5607b927c4774ccf8aa52aa2b1f66a0749cf932076cda97280dfb" exitCode=0 Feb 20 08:31:49 crc kubenswrapper[4948]: I0220 08:31:49.296811 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" event={"ID":"5864a68d-650e-4bcf-b705-619c0f27445b","Type":"ContainerDied","Data":"bb15461fc1a5607b927c4774ccf8aa52aa2b1f66a0749cf932076cda97280dfb"} Feb 20 08:31:49 crc kubenswrapper[4948]: I0220 08:31:49.722707 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:31:49 crc kubenswrapper[4948]: E0220 08:31:49.723130 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:31:50 crc kubenswrapper[4948]: I0220 08:31:50.842671 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:31:50 crc kubenswrapper[4948]: I0220 08:31:50.920638 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgbgc\" (UniqueName: \"kubernetes.io/projected/5864a68d-650e-4bcf-b705-619c0f27445b-kube-api-access-lgbgc\") pod \"5864a68d-650e-4bcf-b705-619c0f27445b\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " Feb 20 08:31:50 crc kubenswrapper[4948]: I0220 08:31:50.921145 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-inventory\") pod \"5864a68d-650e-4bcf-b705-619c0f27445b\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " Feb 20 08:31:50 crc kubenswrapper[4948]: I0220 08:31:50.921490 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-ssh-key-openstack-edpm-ipam\") pod \"5864a68d-650e-4bcf-b705-619c0f27445b\" (UID: \"5864a68d-650e-4bcf-b705-619c0f27445b\") " Feb 20 08:31:50 crc kubenswrapper[4948]: I0220 08:31:50.926098 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5864a68d-650e-4bcf-b705-619c0f27445b-kube-api-access-lgbgc" (OuterVolumeSpecName: "kube-api-access-lgbgc") pod "5864a68d-650e-4bcf-b705-619c0f27445b" (UID: "5864a68d-650e-4bcf-b705-619c0f27445b"). InnerVolumeSpecName "kube-api-access-lgbgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:31:50 crc kubenswrapper[4948]: I0220 08:31:50.951064 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5864a68d-650e-4bcf-b705-619c0f27445b" (UID: "5864a68d-650e-4bcf-b705-619c0f27445b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:31:50 crc kubenswrapper[4948]: I0220 08:31:50.952303 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-inventory" (OuterVolumeSpecName: "inventory") pod "5864a68d-650e-4bcf-b705-619c0f27445b" (UID: "5864a68d-650e-4bcf-b705-619c0f27445b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.024410 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgbgc\" (UniqueName: \"kubernetes.io/projected/5864a68d-650e-4bcf-b705-619c0f27445b-kube-api-access-lgbgc\") on node \"crc\" DevicePath \"\"" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.024440 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.024450 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5864a68d-650e-4bcf-b705-619c0f27445b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.318437 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" event={"ID":"5864a68d-650e-4bcf-b705-619c0f27445b","Type":"ContainerDied","Data":"3ff23193cbb1aafb8b5d373010ecf579e6350e550af86a1ff2fe999d6f3221b3"} Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.318782 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ff23193cbb1aafb8b5d373010ecf579e6350e550af86a1ff2fe999d6f3221b3" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.318529 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-mhkll" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.432496 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx"] Feb 20 08:31:51 crc kubenswrapper[4948]: E0220 08:31:51.433015 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5864a68d-650e-4bcf-b705-619c0f27445b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.433038 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5864a68d-650e-4bcf-b705-619c0f27445b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.433301 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="5864a68d-650e-4bcf-b705-619c0f27445b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.434402 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.439547 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.439847 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.440035 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.440155 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx"] Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.440198 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.534929 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.535141 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnztp\" (UniqueName: \"kubernetes.io/projected/bf3501bf-8f1e-4529-91a8-6be83eda4158-kube-api-access-qnztp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.535214 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.637891 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnztp\" (UniqueName: \"kubernetes.io/projected/bf3501bf-8f1e-4529-91a8-6be83eda4158-kube-api-access-qnztp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.638013 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.638306 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.642410 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.645026 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.656659 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnztp\" (UniqueName: \"kubernetes.io/projected/bf3501bf-8f1e-4529-91a8-6be83eda4158-kube-api-access-qnztp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.760594 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:31:51 crc kubenswrapper[4948]: I0220 08:31:51.769411 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:31:52 crc kubenswrapper[4948]: I0220 08:31:52.354118 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx"] Feb 20 08:31:52 crc kubenswrapper[4948]: I0220 08:31:52.816709 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:31:53 crc kubenswrapper[4948]: I0220 08:31:53.335211 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" event={"ID":"bf3501bf-8f1e-4529-91a8-6be83eda4158","Type":"ContainerStarted","Data":"91d95d8f8184e4a5f35617935d58a284d64d7340d9543fd4f2b969143f1e8c7b"} Feb 20 08:31:53 crc kubenswrapper[4948]: I0220 08:31:53.335306 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" event={"ID":"bf3501bf-8f1e-4529-91a8-6be83eda4158","Type":"ContainerStarted","Data":"64a64af0744255f035f713b739bbbc9dbd942f32b093f55924592b6aeb394758"} Feb 20 08:31:53 crc kubenswrapper[4948]: I0220 08:31:53.350343 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" podStartSLOduration=1.899962839 podStartE2EDuration="2.350326625s" podCreationTimestamp="2026-02-20 08:31:51 +0000 UTC" firstStartedPulling="2026-02-20 08:31:52.360450283 +0000 UTC m=+1561.334945103" lastFinishedPulling="2026-02-20 08:31:52.810814029 +0000 UTC m=+1561.785308889" observedRunningTime="2026-02-20 08:31:53.349018173 +0000 UTC m=+1562.323513003" watchObservedRunningTime="2026-02-20 08:31:53.350326625 +0000 UTC m=+1562.324821445" Feb 20 08:31:55 crc kubenswrapper[4948]: I0220 08:31:55.345838 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:55 crc kubenswrapper[4948]: I0220 08:31:55.438814 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:55 crc kubenswrapper[4948]: I0220 08:31:55.599752 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bdphs"] Feb 20 08:31:57 crc kubenswrapper[4948]: I0220 08:31:57.383663 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bdphs" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="registry-server" containerID="cri-o://37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529" gracePeriod=2 Feb 20 08:31:57 crc kubenswrapper[4948]: I0220 08:31:57.954585 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.072844 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l4qt\" (UniqueName: \"kubernetes.io/projected/060989bf-ef06-4060-bc25-915180f42e5b-kube-api-access-4l4qt\") pod \"060989bf-ef06-4060-bc25-915180f42e5b\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.073099 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-catalog-content\") pod \"060989bf-ef06-4060-bc25-915180f42e5b\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.073248 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-utilities\") pod \"060989bf-ef06-4060-bc25-915180f42e5b\" (UID: \"060989bf-ef06-4060-bc25-915180f42e5b\") " Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.074683 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-utilities" (OuterVolumeSpecName: "utilities") pod "060989bf-ef06-4060-bc25-915180f42e5b" (UID: "060989bf-ef06-4060-bc25-915180f42e5b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.087958 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/060989bf-ef06-4060-bc25-915180f42e5b-kube-api-access-4l4qt" (OuterVolumeSpecName: "kube-api-access-4l4qt") pod "060989bf-ef06-4060-bc25-915180f42e5b" (UID: "060989bf-ef06-4060-bc25-915180f42e5b"). InnerVolumeSpecName "kube-api-access-4l4qt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.174948 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.175237 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l4qt\" (UniqueName: \"kubernetes.io/projected/060989bf-ef06-4060-bc25-915180f42e5b-kube-api-access-4l4qt\") on node \"crc\" DevicePath \"\"" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.191165 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "060989bf-ef06-4060-bc25-915180f42e5b" (UID: "060989bf-ef06-4060-bc25-915180f42e5b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.276646 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/060989bf-ef06-4060-bc25-915180f42e5b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.396524 4948 generic.go:334] "Generic (PLEG): container finished" podID="060989bf-ef06-4060-bc25-915180f42e5b" containerID="37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529" exitCode=0 Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.396650 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdphs" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.401113 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdphs" event={"ID":"060989bf-ef06-4060-bc25-915180f42e5b","Type":"ContainerDied","Data":"37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529"} Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.401208 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdphs" event={"ID":"060989bf-ef06-4060-bc25-915180f42e5b","Type":"ContainerDied","Data":"53735cceaab8db6cbf509cf3d7a338fd07617ca1b94b6981218603729cf70256"} Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.401242 4948 scope.go:117] "RemoveContainer" containerID="37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.423202 4948 scope.go:117] "RemoveContainer" containerID="7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.440024 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bdphs"] Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.449474 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bdphs"] Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.467923 4948 scope.go:117] "RemoveContainer" containerID="cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.539144 4948 scope.go:117] "RemoveContainer" containerID="37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529" Feb 20 08:31:58 crc kubenswrapper[4948]: E0220 08:31:58.540360 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529\": container with ID starting with 37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529 not found: ID does not exist" containerID="37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.540390 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529"} err="failed to get container status \"37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529\": rpc error: code = NotFound desc = could not find container \"37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529\": container with ID starting with 37c7e4d25f9127067f12ac87e6ea1ca5dbb8e45ce19e01d5f213b5bb1cd40529 not found: ID does not exist" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.540410 4948 scope.go:117] "RemoveContainer" containerID="7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966" Feb 20 08:31:58 crc kubenswrapper[4948]: E0220 08:31:58.540695 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966\": container with ID starting with 7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966 not found: ID does not exist" containerID="7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.540715 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966"} err="failed to get container status \"7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966\": rpc error: code = NotFound desc = could not find container \"7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966\": container with ID starting with 7cfda758c2e1d88101427b25367b79d9870e89cc2808d38e69b1e5dfa5780966 not found: ID does not exist" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.540727 4948 scope.go:117] "RemoveContainer" containerID="cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1" Feb 20 08:31:58 crc kubenswrapper[4948]: E0220 08:31:58.540990 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1\": container with ID starting with cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1 not found: ID does not exist" containerID="cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.541010 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1"} err="failed to get container status \"cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1\": rpc error: code = NotFound desc = could not find container \"cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1\": container with ID starting with cafef04c966377e665f78102a5914bb6441b7f707970965b08dc70cfb2cf82a1 not found: ID does not exist" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.887000 4948 scope.go:117] "RemoveContainer" containerID="f4eb6ffc3a340bf3fb020244f7c41526c57d80da4f478c0885a3916041b2da30" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.915904 4948 scope.go:117] "RemoveContainer" containerID="67a7241e5369f13541e0b84997b90ee0fd96a87f376c9c419e0e637fa48068da" Feb 20 08:31:58 crc kubenswrapper[4948]: I0220 08:31:58.972338 4948 scope.go:117] "RemoveContainer" containerID="279cb01f22f256d8185916f133561b9dbb69d1fb0ef3e531faed5ee14470b9d2" Feb 20 08:31:59 crc kubenswrapper[4948]: I0220 08:31:59.034929 4948 scope.go:117] "RemoveContainer" containerID="043fb8ebd535abc3ac64988e423e5b76f657b01785d57b024ae113afd44b4bf7" Feb 20 08:31:59 crc kubenswrapper[4948]: I0220 08:31:59.083922 4948 scope.go:117] "RemoveContainer" containerID="eb75184d7c11178b70689bbef778a20eca59576d9cf6e2826963a0598aa904a4" Feb 20 08:31:59 crc kubenswrapper[4948]: I0220 08:31:59.116349 4948 scope.go:117] "RemoveContainer" containerID="8e0c0f098f3973e04052c6dc5c4da1ff436370579d9b1ea6857d9e0c6bed683e" Feb 20 08:31:59 crc kubenswrapper[4948]: I0220 08:31:59.748548 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="060989bf-ef06-4060-bc25-915180f42e5b" path="/var/lib/kubelet/pods/060989bf-ef06-4060-bc25-915180f42e5b/volumes" Feb 20 08:32:00 crc kubenswrapper[4948]: I0220 08:32:00.044180 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-a86b-account-create-update-mspp7"] Feb 20 08:32:00 crc kubenswrapper[4948]: I0220 08:32:00.064553 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-97j89"] Feb 20 08:32:00 crc kubenswrapper[4948]: I0220 08:32:00.075931 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-4djkr"] Feb 20 08:32:00 crc kubenswrapper[4948]: I0220 08:32:00.086152 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-a86b-account-create-update-mspp7"] Feb 20 08:32:00 crc kubenswrapper[4948]: I0220 08:32:00.096064 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-97j89"] Feb 20 08:32:00 crc kubenswrapper[4948]: I0220 08:32:00.102904 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-4djkr"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.047766 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-e6d0-account-create-update-cdxd5"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.070049 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-w4bzp"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.089314 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-w4bzp"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.099364 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-e6d0-account-create-update-cdxd5"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.107785 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-0b2f-account-create-update-2f74q"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.115484 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-zvzvk"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.123401 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-zvzvk"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.130817 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-0b2f-account-create-update-2f74q"] Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.743731 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a739e4d-8960-46f3-8c5d-2373b59165dd" path="/var/lib/kubelet/pods/0a739e4d-8960-46f3-8c5d-2373b59165dd/volumes" Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.746533 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10eedd01-3f52-4d44-981d-27bc16af186b" path="/var/lib/kubelet/pods/10eedd01-3f52-4d44-981d-27bc16af186b/volumes" Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.748126 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12f1f665-7a04-4794-a727-cc5c060a1933" path="/var/lib/kubelet/pods/12f1f665-7a04-4794-a727-cc5c060a1933/volumes" Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.749568 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ced4ac5-e057-476a-908a-f573a8cf997b" path="/var/lib/kubelet/pods/4ced4ac5-e057-476a-908a-f573a8cf997b/volumes" Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.752240 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="501e504c-7910-47cb-a181-812a0d72f2db" path="/var/lib/kubelet/pods/501e504c-7910-47cb-a181-812a0d72f2db/volumes" Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.754700 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6d068f3-210b-41e8-b90a-5406759c0606" path="/var/lib/kubelet/pods/b6d068f3-210b-41e8-b90a-5406759c0606/volumes" Feb 20 08:32:01 crc kubenswrapper[4948]: I0220 08:32:01.756880 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3f86085-113f-408e-b43c-a499e2fe8b90" path="/var/lib/kubelet/pods/d3f86085-113f-408e-b43c-a499e2fe8b90/volumes" Feb 20 08:32:04 crc kubenswrapper[4948]: I0220 08:32:04.724532 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:32:04 crc kubenswrapper[4948]: E0220 08:32:04.725418 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:32:12 crc kubenswrapper[4948]: I0220 08:32:12.063107 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-pkgzx"] Feb 20 08:32:12 crc kubenswrapper[4948]: I0220 08:32:12.077308 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-pkgzx"] Feb 20 08:32:13 crc kubenswrapper[4948]: I0220 08:32:13.036154 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-gz5j7"] Feb 20 08:32:13 crc kubenswrapper[4948]: I0220 08:32:13.047571 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-gz5j7"] Feb 20 08:32:13 crc kubenswrapper[4948]: I0220 08:32:13.756751 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65735728-e734-40de-a185-9caa1db9a47d" path="/var/lib/kubelet/pods/65735728-e734-40de-a185-9caa1db9a47d/volumes" Feb 20 08:32:13 crc kubenswrapper[4948]: I0220 08:32:13.759151 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9370de8-8b72-4819-af79-4a9ee1fb6777" path="/var/lib/kubelet/pods/b9370de8-8b72-4819-af79-4a9ee1fb6777/volumes" Feb 20 08:32:17 crc kubenswrapper[4948]: I0220 08:32:17.723505 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:32:17 crc kubenswrapper[4948]: E0220 08:32:17.724689 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:32:30 crc kubenswrapper[4948]: I0220 08:32:30.723247 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:32:30 crc kubenswrapper[4948]: E0220 08:32:30.724285 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:32:41 crc kubenswrapper[4948]: I0220 08:32:41.060853 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-g5zxh"] Feb 20 08:32:41 crc kubenswrapper[4948]: I0220 08:32:41.073044 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-g5zxh"] Feb 20 08:32:41 crc kubenswrapper[4948]: I0220 08:32:41.745863 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb362d0e-68b4-4d48-885c-a6894ac71e6e" path="/var/lib/kubelet/pods/bb362d0e-68b4-4d48-885c-a6894ac71e6e/volumes" Feb 20 08:32:43 crc kubenswrapper[4948]: I0220 08:32:43.723862 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:32:43 crc kubenswrapper[4948]: E0220 08:32:43.724478 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:32:43 crc kubenswrapper[4948]: I0220 08:32:43.959617 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wfjj9"] Feb 20 08:32:43 crc kubenswrapper[4948]: E0220 08:32:43.960297 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="registry-server" Feb 20 08:32:43 crc kubenswrapper[4948]: I0220 08:32:43.960329 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="registry-server" Feb 20 08:32:43 crc kubenswrapper[4948]: E0220 08:32:43.960404 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="extract-utilities" Feb 20 08:32:43 crc kubenswrapper[4948]: I0220 08:32:43.960419 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="extract-utilities" Feb 20 08:32:43 crc kubenswrapper[4948]: E0220 08:32:43.960451 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="extract-content" Feb 20 08:32:43 crc kubenswrapper[4948]: I0220 08:32:43.960464 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="extract-content" Feb 20 08:32:43 crc kubenswrapper[4948]: I0220 08:32:43.960812 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="060989bf-ef06-4060-bc25-915180f42e5b" containerName="registry-server" Feb 20 08:32:43 crc kubenswrapper[4948]: I0220 08:32:43.963553 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:43 crc kubenswrapper[4948]: I0220 08:32:43.976177 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wfjj9"] Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.038045 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng6rg\" (UniqueName: \"kubernetes.io/projected/127d52c1-3c58-424b-aa24-d71a40715228-kube-api-access-ng6rg\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.038143 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-catalog-content\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.038185 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-utilities\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.140508 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng6rg\" (UniqueName: \"kubernetes.io/projected/127d52c1-3c58-424b-aa24-d71a40715228-kube-api-access-ng6rg\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.141110 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-catalog-content\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.141647 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-catalog-content\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.141808 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-utilities\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.142180 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-utilities\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.162953 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng6rg\" (UniqueName: \"kubernetes.io/projected/127d52c1-3c58-424b-aa24-d71a40715228-kube-api-access-ng6rg\") pod \"certified-operators-wfjj9\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.298192 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:44 crc kubenswrapper[4948]: I0220 08:32:44.792859 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wfjj9"] Feb 20 08:32:45 crc kubenswrapper[4948]: I0220 08:32:45.009183 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfjj9" event={"ID":"127d52c1-3c58-424b-aa24-d71a40715228","Type":"ContainerStarted","Data":"105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed"} Feb 20 08:32:45 crc kubenswrapper[4948]: I0220 08:32:45.009501 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfjj9" event={"ID":"127d52c1-3c58-424b-aa24-d71a40715228","Type":"ContainerStarted","Data":"162bd0680e1ceb7df91470f556615acedd676fe4c2560f55f65606b3bd82a0ad"} Feb 20 08:32:46 crc kubenswrapper[4948]: I0220 08:32:46.025432 4948 generic.go:334] "Generic (PLEG): container finished" podID="127d52c1-3c58-424b-aa24-d71a40715228" containerID="105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed" exitCode=0 Feb 20 08:32:46 crc kubenswrapper[4948]: I0220 08:32:46.025506 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfjj9" event={"ID":"127d52c1-3c58-424b-aa24-d71a40715228","Type":"ContainerDied","Data":"105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed"} Feb 20 08:32:46 crc kubenswrapper[4948]: I0220 08:32:46.029149 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:32:47 crc kubenswrapper[4948]: I0220 08:32:47.030189 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-6mz7x"] Feb 20 08:32:47 crc kubenswrapper[4948]: I0220 08:32:47.038077 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfjj9" event={"ID":"127d52c1-3c58-424b-aa24-d71a40715228","Type":"ContainerStarted","Data":"88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5"} Feb 20 08:32:47 crc kubenswrapper[4948]: I0220 08:32:47.039242 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-6mz7x"] Feb 20 08:32:47 crc kubenswrapper[4948]: I0220 08:32:47.747463 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1227b752-8a3c-4c01-91bf-0662c1b1a231" path="/var/lib/kubelet/pods/1227b752-8a3c-4c01-91bf-0662c1b1a231/volumes" Feb 20 08:32:48 crc kubenswrapper[4948]: I0220 08:32:48.053294 4948 generic.go:334] "Generic (PLEG): container finished" podID="127d52c1-3c58-424b-aa24-d71a40715228" containerID="88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5" exitCode=0 Feb 20 08:32:48 crc kubenswrapper[4948]: I0220 08:32:48.053358 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfjj9" event={"ID":"127d52c1-3c58-424b-aa24-d71a40715228","Type":"ContainerDied","Data":"88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5"} Feb 20 08:32:49 crc kubenswrapper[4948]: I0220 08:32:49.066244 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfjj9" event={"ID":"127d52c1-3c58-424b-aa24-d71a40715228","Type":"ContainerStarted","Data":"058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31"} Feb 20 08:32:49 crc kubenswrapper[4948]: I0220 08:32:49.106419 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wfjj9" podStartSLOduration=3.695804652 podStartE2EDuration="6.106395068s" podCreationTimestamp="2026-02-20 08:32:43 +0000 UTC" firstStartedPulling="2026-02-20 08:32:46.028561749 +0000 UTC m=+1615.003056599" lastFinishedPulling="2026-02-20 08:32:48.439152195 +0000 UTC m=+1617.413647015" observedRunningTime="2026-02-20 08:32:49.099582089 +0000 UTC m=+1618.074076909" watchObservedRunningTime="2026-02-20 08:32:49.106395068 +0000 UTC m=+1618.080889908" Feb 20 08:32:50 crc kubenswrapper[4948]: I0220 08:32:50.061325 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-v29fx"] Feb 20 08:32:50 crc kubenswrapper[4948]: I0220 08:32:50.077529 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-v29fx"] Feb 20 08:32:51 crc kubenswrapper[4948]: I0220 08:32:51.738726 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d84a71b6-7503-4150-a5c4-4579b08a669a" path="/var/lib/kubelet/pods/d84a71b6-7503-4150-a5c4-4579b08a669a/volumes" Feb 20 08:32:54 crc kubenswrapper[4948]: I0220 08:32:54.300461 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:54 crc kubenswrapper[4948]: I0220 08:32:54.300545 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:54 crc kubenswrapper[4948]: I0220 08:32:54.392448 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:55 crc kubenswrapper[4948]: I0220 08:32:55.228922 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:55 crc kubenswrapper[4948]: I0220 08:32:55.309610 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wfjj9"] Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.166778 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wfjj9" podUID="127d52c1-3c58-424b-aa24-d71a40715228" containerName="registry-server" containerID="cri-o://058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31" gracePeriod=2 Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.723142 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:32:57 crc kubenswrapper[4948]: E0220 08:32:57.723749 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.735128 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.854092 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ng6rg\" (UniqueName: \"kubernetes.io/projected/127d52c1-3c58-424b-aa24-d71a40715228-kube-api-access-ng6rg\") pod \"127d52c1-3c58-424b-aa24-d71a40715228\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.854494 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-catalog-content\") pod \"127d52c1-3c58-424b-aa24-d71a40715228\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.854599 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-utilities\") pod \"127d52c1-3c58-424b-aa24-d71a40715228\" (UID: \"127d52c1-3c58-424b-aa24-d71a40715228\") " Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.857961 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-utilities" (OuterVolumeSpecName: "utilities") pod "127d52c1-3c58-424b-aa24-d71a40715228" (UID: "127d52c1-3c58-424b-aa24-d71a40715228"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.864053 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/127d52c1-3c58-424b-aa24-d71a40715228-kube-api-access-ng6rg" (OuterVolumeSpecName: "kube-api-access-ng6rg") pod "127d52c1-3c58-424b-aa24-d71a40715228" (UID: "127d52c1-3c58-424b-aa24-d71a40715228"). InnerVolumeSpecName "kube-api-access-ng6rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.960689 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:32:57 crc kubenswrapper[4948]: I0220 08:32:57.960749 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ng6rg\" (UniqueName: \"kubernetes.io/projected/127d52c1-3c58-424b-aa24-d71a40715228-kube-api-access-ng6rg\") on node \"crc\" DevicePath \"\"" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.010920 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "127d52c1-3c58-424b-aa24-d71a40715228" (UID: "127d52c1-3c58-424b-aa24-d71a40715228"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.062304 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/127d52c1-3c58-424b-aa24-d71a40715228-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.179830 4948 generic.go:334] "Generic (PLEG): container finished" podID="127d52c1-3c58-424b-aa24-d71a40715228" containerID="058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31" exitCode=0 Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.179881 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfjj9" event={"ID":"127d52c1-3c58-424b-aa24-d71a40715228","Type":"ContainerDied","Data":"058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31"} Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.179912 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfjj9" event={"ID":"127d52c1-3c58-424b-aa24-d71a40715228","Type":"ContainerDied","Data":"162bd0680e1ceb7df91470f556615acedd676fe4c2560f55f65606b3bd82a0ad"} Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.179934 4948 scope.go:117] "RemoveContainer" containerID="058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.180144 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfjj9" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.228109 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wfjj9"] Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.236537 4948 scope.go:117] "RemoveContainer" containerID="88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.242662 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wfjj9"] Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.274530 4948 scope.go:117] "RemoveContainer" containerID="105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.342643 4948 scope.go:117] "RemoveContainer" containerID="058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31" Feb 20 08:32:58 crc kubenswrapper[4948]: E0220 08:32:58.343234 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31\": container with ID starting with 058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31 not found: ID does not exist" containerID="058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.343276 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31"} err="failed to get container status \"058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31\": rpc error: code = NotFound desc = could not find container \"058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31\": container with ID starting with 058114f0c0b1e6905b0749e5f435db738c2b0ddf4ed180217b4ddd8152ed9e31 not found: ID does not exist" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.343302 4948 scope.go:117] "RemoveContainer" containerID="88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5" Feb 20 08:32:58 crc kubenswrapper[4948]: E0220 08:32:58.343720 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5\": container with ID starting with 88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5 not found: ID does not exist" containerID="88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.343887 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5"} err="failed to get container status \"88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5\": rpc error: code = NotFound desc = could not find container \"88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5\": container with ID starting with 88f25f8239c60da55713318b991058f1de64af08883114ae420d1004b33150e5 not found: ID does not exist" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.344023 4948 scope.go:117] "RemoveContainer" containerID="105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed" Feb 20 08:32:58 crc kubenswrapper[4948]: E0220 08:32:58.344458 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed\": container with ID starting with 105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed not found: ID does not exist" containerID="105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed" Feb 20 08:32:58 crc kubenswrapper[4948]: I0220 08:32:58.344576 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed"} err="failed to get container status \"105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed\": rpc error: code = NotFound desc = could not find container \"105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed\": container with ID starting with 105e4d157fc4d190d06fdac23834273f0c6149fd0638550ef4ca602d493f78ed not found: ID does not exist" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.197647 4948 generic.go:334] "Generic (PLEG): container finished" podID="bf3501bf-8f1e-4529-91a8-6be83eda4158" containerID="91d95d8f8184e4a5f35617935d58a284d64d7340d9543fd4f2b969143f1e8c7b" exitCode=0 Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.197733 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" event={"ID":"bf3501bf-8f1e-4529-91a8-6be83eda4158","Type":"ContainerDied","Data":"91d95d8f8184e4a5f35617935d58a284d64d7340d9543fd4f2b969143f1e8c7b"} Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.326685 4948 scope.go:117] "RemoveContainer" containerID="6457842f65c42b8313a38ca922ee61a97a4823ed4c56040981289a5c5705cd46" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.365367 4948 scope.go:117] "RemoveContainer" containerID="f8ab1b0def6ea5b1022acdb9c87659d2c1c20a66a736d1e630b89d097c36d6ef" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.438707 4948 scope.go:117] "RemoveContainer" containerID="87c2cdb1a2b7f5c38a01dd7f884ce5938563a33612ebaf51f20ceac5532f97f2" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.519780 4948 scope.go:117] "RemoveContainer" containerID="fb257235a0241c571d44f0c21ce5fac86cd6393e8a1250cb5e78d4e1b599fb51" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.553943 4948 scope.go:117] "RemoveContainer" containerID="1fb571973e057c8519a5db05840734117d94ed09a1e521d57cf4a92fa88cbcfb" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.614812 4948 scope.go:117] "RemoveContainer" containerID="7e3cc540d00613674fe548789ab2ccc71a06c5714f420cedea97409a8254b14f" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.640400 4948 scope.go:117] "RemoveContainer" containerID="c565785e94684cef9e257c737e709e95f61b9c29c0e4f1d8e37358207acddb4c" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.672123 4948 scope.go:117] "RemoveContainer" containerID="b8d5cd74c5cbd29dae34c7b1cedbde475f451d2de997a59b511db6cc61ba0a29" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.697613 4948 scope.go:117] "RemoveContainer" containerID="cddaaa0c08e72fa702328c57550853b538c4537573467ac0ebd0761e1fcbc30c" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.720067 4948 scope.go:117] "RemoveContainer" containerID="cb7241154041475439806eb631212532adb6e6775adf70e62bdf5cf4b6d19276" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.741746 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="127d52c1-3c58-424b-aa24-d71a40715228" path="/var/lib/kubelet/pods/127d52c1-3c58-424b-aa24-d71a40715228/volumes" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.757837 4948 scope.go:117] "RemoveContainer" containerID="72b50822f4c48e3166b0066730c258d864d6094e98d9d9617ca0abdc1f36ae38" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.797425 4948 scope.go:117] "RemoveContainer" containerID="a11f29a4f832ccc8dd04f52fcb05e1f418934455b1143a9d46f2d5ad1bce9425" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.826276 4948 scope.go:117] "RemoveContainer" containerID="ede14cf7eee35beb81bccaf96c5abbd270f76ba98b65d67579bbb482b130f45e" Feb 20 08:32:59 crc kubenswrapper[4948]: I0220 08:32:59.847415 4948 scope.go:117] "RemoveContainer" containerID="ddcd74111702fbb6556b1046c04ec320ea3aecaa381a53da7a6aa1bc0c3d6fd7" Feb 20 08:33:00 crc kubenswrapper[4948]: I0220 08:33:00.860103 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:33:00 crc kubenswrapper[4948]: I0220 08:33:00.924819 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-inventory\") pod \"bf3501bf-8f1e-4529-91a8-6be83eda4158\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " Feb 20 08:33:00 crc kubenswrapper[4948]: I0220 08:33:00.925046 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnztp\" (UniqueName: \"kubernetes.io/projected/bf3501bf-8f1e-4529-91a8-6be83eda4158-kube-api-access-qnztp\") pod \"bf3501bf-8f1e-4529-91a8-6be83eda4158\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " Feb 20 08:33:00 crc kubenswrapper[4948]: I0220 08:33:00.925344 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-ssh-key-openstack-edpm-ipam\") pod \"bf3501bf-8f1e-4529-91a8-6be83eda4158\" (UID: \"bf3501bf-8f1e-4529-91a8-6be83eda4158\") " Feb 20 08:33:00 crc kubenswrapper[4948]: I0220 08:33:00.930925 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf3501bf-8f1e-4529-91a8-6be83eda4158-kube-api-access-qnztp" (OuterVolumeSpecName: "kube-api-access-qnztp") pod "bf3501bf-8f1e-4529-91a8-6be83eda4158" (UID: "bf3501bf-8f1e-4529-91a8-6be83eda4158"). InnerVolumeSpecName "kube-api-access-qnztp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:33:00 crc kubenswrapper[4948]: I0220 08:33:00.955695 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "bf3501bf-8f1e-4529-91a8-6be83eda4158" (UID: "bf3501bf-8f1e-4529-91a8-6be83eda4158"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:33:00 crc kubenswrapper[4948]: I0220 08:33:00.971417 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-inventory" (OuterVolumeSpecName: "inventory") pod "bf3501bf-8f1e-4529-91a8-6be83eda4158" (UID: "bf3501bf-8f1e-4529-91a8-6be83eda4158"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.027820 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.027875 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf3501bf-8f1e-4529-91a8-6be83eda4158-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.027895 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnztp\" (UniqueName: \"kubernetes.io/projected/bf3501bf-8f1e-4529-91a8-6be83eda4158-kube-api-access-qnztp\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.238906 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" event={"ID":"bf3501bf-8f1e-4529-91a8-6be83eda4158","Type":"ContainerDied","Data":"64a64af0744255f035f713b739bbbc9dbd942f32b093f55924592b6aeb394758"} Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.240180 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64a64af0744255f035f713b739bbbc9dbd942f32b093f55924592b6aeb394758" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.239002 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.542434 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d"] Feb 20 08:33:01 crc kubenswrapper[4948]: E0220 08:33:01.542799 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="127d52c1-3c58-424b-aa24-d71a40715228" containerName="extract-utilities" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.542816 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="127d52c1-3c58-424b-aa24-d71a40715228" containerName="extract-utilities" Feb 20 08:33:01 crc kubenswrapper[4948]: E0220 08:33:01.542833 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf3501bf-8f1e-4529-91a8-6be83eda4158" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.542841 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf3501bf-8f1e-4529-91a8-6be83eda4158" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:01 crc kubenswrapper[4948]: E0220 08:33:01.542866 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="127d52c1-3c58-424b-aa24-d71a40715228" containerName="extract-content" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.542872 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="127d52c1-3c58-424b-aa24-d71a40715228" containerName="extract-content" Feb 20 08:33:01 crc kubenswrapper[4948]: E0220 08:33:01.542884 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="127d52c1-3c58-424b-aa24-d71a40715228" containerName="registry-server" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.542890 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="127d52c1-3c58-424b-aa24-d71a40715228" containerName="registry-server" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.543079 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="127d52c1-3c58-424b-aa24-d71a40715228" containerName="registry-server" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.543094 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf3501bf-8f1e-4529-91a8-6be83eda4158" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.543705 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.546852 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.546864 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.547532 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.551795 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.574374 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d"] Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.639823 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.639902 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.640297 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfwmf\" (UniqueName: \"kubernetes.io/projected/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-kube-api-access-vfwmf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.742176 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.742264 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.742462 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfwmf\" (UniqueName: \"kubernetes.io/projected/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-kube-api-access-vfwmf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.749431 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.750035 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.772884 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfwmf\" (UniqueName: \"kubernetes.io/projected/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-kube-api-access-vfwmf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lf98d\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:01 crc kubenswrapper[4948]: I0220 08:33:01.863514 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:02 crc kubenswrapper[4948]: I0220 08:33:02.424340 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d"] Feb 20 08:33:03 crc kubenswrapper[4948]: I0220 08:33:03.264466 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" event={"ID":"2e05be38-f4ab-415f-a71d-9e5233cf0ea7","Type":"ContainerStarted","Data":"345677fd825b78a8c9a15733dc5504746aec837d4a4da61dd8b843aa56c4878f"} Feb 20 08:33:03 crc kubenswrapper[4948]: I0220 08:33:03.265091 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" event={"ID":"2e05be38-f4ab-415f-a71d-9e5233cf0ea7","Type":"ContainerStarted","Data":"3eb7fcb6fef9877434c717228e6f27c38c4af180ee08b3677022845e3cf3f71d"} Feb 20 08:33:03 crc kubenswrapper[4948]: I0220 08:33:03.291785 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" podStartSLOduration=1.8341930990000002 podStartE2EDuration="2.291766439s" podCreationTimestamp="2026-02-20 08:33:01 +0000 UTC" firstStartedPulling="2026-02-20 08:33:02.439579229 +0000 UTC m=+1631.414074049" lastFinishedPulling="2026-02-20 08:33:02.897152529 +0000 UTC m=+1631.871647389" observedRunningTime="2026-02-20 08:33:03.281892685 +0000 UTC m=+1632.256387525" watchObservedRunningTime="2026-02-20 08:33:03.291766439 +0000 UTC m=+1632.266261259" Feb 20 08:33:04 crc kubenswrapper[4948]: I0220 08:33:04.050397 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-xt2zq"] Feb 20 08:33:04 crc kubenswrapper[4948]: I0220 08:33:04.061342 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-xt2zq"] Feb 20 08:33:05 crc kubenswrapper[4948]: I0220 08:33:05.738442 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f706b741-86f9-49a3-95d5-85ee62eb3668" path="/var/lib/kubelet/pods/f706b741-86f9-49a3-95d5-85ee62eb3668/volumes" Feb 20 08:33:06 crc kubenswrapper[4948]: I0220 08:33:06.028657 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-c7zzv"] Feb 20 08:33:06 crc kubenswrapper[4948]: I0220 08:33:06.042626 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-c7zzv"] Feb 20 08:33:07 crc kubenswrapper[4948]: I0220 08:33:07.746089 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1f108e1-27bb-433e-b7e1-f9cc7f778182" path="/var/lib/kubelet/pods/f1f108e1-27bb-433e-b7e1-f9cc7f778182/volumes" Feb 20 08:33:08 crc kubenswrapper[4948]: I0220 08:33:08.337501 4948 generic.go:334] "Generic (PLEG): container finished" podID="2e05be38-f4ab-415f-a71d-9e5233cf0ea7" containerID="345677fd825b78a8c9a15733dc5504746aec837d4a4da61dd8b843aa56c4878f" exitCode=0 Feb 20 08:33:08 crc kubenswrapper[4948]: I0220 08:33:08.337625 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" event={"ID":"2e05be38-f4ab-415f-a71d-9e5233cf0ea7","Type":"ContainerDied","Data":"345677fd825b78a8c9a15733dc5504746aec837d4a4da61dd8b843aa56c4878f"} Feb 20 08:33:09 crc kubenswrapper[4948]: I0220 08:33:09.882849 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.065247 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-inventory\") pod \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.065509 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-ssh-key-openstack-edpm-ipam\") pod \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.065576 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfwmf\" (UniqueName: \"kubernetes.io/projected/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-kube-api-access-vfwmf\") pod \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\" (UID: \"2e05be38-f4ab-415f-a71d-9e5233cf0ea7\") " Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.073768 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-kube-api-access-vfwmf" (OuterVolumeSpecName: "kube-api-access-vfwmf") pod "2e05be38-f4ab-415f-a71d-9e5233cf0ea7" (UID: "2e05be38-f4ab-415f-a71d-9e5233cf0ea7"). InnerVolumeSpecName "kube-api-access-vfwmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.094400 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-inventory" (OuterVolumeSpecName: "inventory") pod "2e05be38-f4ab-415f-a71d-9e5233cf0ea7" (UID: "2e05be38-f4ab-415f-a71d-9e5233cf0ea7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.106385 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2e05be38-f4ab-415f-a71d-9e5233cf0ea7" (UID: "2e05be38-f4ab-415f-a71d-9e5233cf0ea7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.168203 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.168242 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfwmf\" (UniqueName: \"kubernetes.io/projected/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-kube-api-access-vfwmf\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.168256 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e05be38-f4ab-415f-a71d-9e5233cf0ea7-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.367167 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" event={"ID":"2e05be38-f4ab-415f-a71d-9e5233cf0ea7","Type":"ContainerDied","Data":"3eb7fcb6fef9877434c717228e6f27c38c4af180ee08b3677022845e3cf3f71d"} Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.367226 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3eb7fcb6fef9877434c717228e6f27c38c4af180ee08b3677022845e3cf3f71d" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.367351 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lf98d" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.586490 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v"] Feb 20 08:33:10 crc kubenswrapper[4948]: E0220 08:33:10.588413 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e05be38-f4ab-415f-a71d-9e5233cf0ea7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.588465 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e05be38-f4ab-415f-a71d-9e5233cf0ea7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.589558 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e05be38-f4ab-415f-a71d-9e5233cf0ea7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.593538 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.600283 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.600367 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.600397 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.601167 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.611849 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v"] Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.677414 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.677753 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxqrc\" (UniqueName: \"kubernetes.io/projected/f42ddd0c-a2ec-414a-812d-f87213a3226d-kube-api-access-dxqrc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.677850 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.779586 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.779714 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.779843 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxqrc\" (UniqueName: \"kubernetes.io/projected/f42ddd0c-a2ec-414a-812d-f87213a3226d-kube-api-access-dxqrc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.785589 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.786901 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.805213 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxqrc\" (UniqueName: \"kubernetes.io/projected/f42ddd0c-a2ec-414a-812d-f87213a3226d-kube-api-access-dxqrc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l55v\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:10 crc kubenswrapper[4948]: I0220 08:33:10.931970 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:11 crc kubenswrapper[4948]: I0220 08:33:11.501608 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v"] Feb 20 08:33:11 crc kubenswrapper[4948]: I0220 08:33:11.729047 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:33:11 crc kubenswrapper[4948]: E0220 08:33:11.729345 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:33:12 crc kubenswrapper[4948]: I0220 08:33:12.393255 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" event={"ID":"f42ddd0c-a2ec-414a-812d-f87213a3226d","Type":"ContainerStarted","Data":"ddcb74ec6513e9050b354f5584993e70b4784d82329c695b8cbcea156ddada02"} Feb 20 08:33:12 crc kubenswrapper[4948]: I0220 08:33:12.394009 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" event={"ID":"f42ddd0c-a2ec-414a-812d-f87213a3226d","Type":"ContainerStarted","Data":"26e225a9996f58185c76eebc01d4b0d42e12e26c4c12d0fd5527a74efcfa0e7f"} Feb 20 08:33:12 crc kubenswrapper[4948]: I0220 08:33:12.420865 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" podStartSLOduration=2.015293287 podStartE2EDuration="2.420844738s" podCreationTimestamp="2026-02-20 08:33:10 +0000 UTC" firstStartedPulling="2026-02-20 08:33:11.50705241 +0000 UTC m=+1640.481547230" lastFinishedPulling="2026-02-20 08:33:11.912603851 +0000 UTC m=+1640.887098681" observedRunningTime="2026-02-20 08:33:12.419385541 +0000 UTC m=+1641.393880361" watchObservedRunningTime="2026-02-20 08:33:12.420844738 +0000 UTC m=+1641.395339558" Feb 20 08:33:22 crc kubenswrapper[4948]: I0220 08:33:22.722757 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:33:22 crc kubenswrapper[4948]: E0220 08:33:22.725554 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.564806 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lvjzf"] Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.567279 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.600761 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lvjzf"] Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.667489 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jfnl\" (UniqueName: \"kubernetes.io/projected/26bee9f7-36e7-4452-be46-a7af14fe0d82-kube-api-access-2jfnl\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.667765 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-utilities\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.667937 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-catalog-content\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.769407 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jfnl\" (UniqueName: \"kubernetes.io/projected/26bee9f7-36e7-4452-be46-a7af14fe0d82-kube-api-access-2jfnl\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.769540 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-utilities\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.769619 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-catalog-content\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.770092 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-catalog-content\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.770090 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-utilities\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.798173 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jfnl\" (UniqueName: \"kubernetes.io/projected/26bee9f7-36e7-4452-be46-a7af14fe0d82-kube-api-access-2jfnl\") pod \"community-operators-lvjzf\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:23 crc kubenswrapper[4948]: I0220 08:33:23.894405 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:24 crc kubenswrapper[4948]: I0220 08:33:24.417039 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lvjzf"] Feb 20 08:33:24 crc kubenswrapper[4948]: I0220 08:33:24.544625 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvjzf" event={"ID":"26bee9f7-36e7-4452-be46-a7af14fe0d82","Type":"ContainerStarted","Data":"d87e55975b4c132e628eaad19e343e2dc1a36778a825ecd32677e5386de184a8"} Feb 20 08:33:24 crc kubenswrapper[4948]: E0220 08:33:24.823961 4948 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26bee9f7_36e7_4452_be46_a7af14fe0d82.slice/crio-conmon-d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a.scope\": RecentStats: unable to find data in memory cache]" Feb 20 08:33:25 crc kubenswrapper[4948]: I0220 08:33:25.559497 4948 generic.go:334] "Generic (PLEG): container finished" podID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerID="d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a" exitCode=0 Feb 20 08:33:25 crc kubenswrapper[4948]: I0220 08:33:25.559621 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvjzf" event={"ID":"26bee9f7-36e7-4452-be46-a7af14fe0d82","Type":"ContainerDied","Data":"d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a"} Feb 20 08:33:26 crc kubenswrapper[4948]: I0220 08:33:26.572659 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvjzf" event={"ID":"26bee9f7-36e7-4452-be46-a7af14fe0d82","Type":"ContainerStarted","Data":"3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c"} Feb 20 08:33:27 crc kubenswrapper[4948]: I0220 08:33:27.582706 4948 generic.go:334] "Generic (PLEG): container finished" podID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerID="3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c" exitCode=0 Feb 20 08:33:27 crc kubenswrapper[4948]: I0220 08:33:27.582745 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvjzf" event={"ID":"26bee9f7-36e7-4452-be46-a7af14fe0d82","Type":"ContainerDied","Data":"3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c"} Feb 20 08:33:28 crc kubenswrapper[4948]: I0220 08:33:28.593448 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvjzf" event={"ID":"26bee9f7-36e7-4452-be46-a7af14fe0d82","Type":"ContainerStarted","Data":"2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d"} Feb 20 08:33:28 crc kubenswrapper[4948]: I0220 08:33:28.612331 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lvjzf" podStartSLOduration=3.184383818 podStartE2EDuration="5.612310293s" podCreationTimestamp="2026-02-20 08:33:23 +0000 UTC" firstStartedPulling="2026-02-20 08:33:25.562612168 +0000 UTC m=+1654.537106988" lastFinishedPulling="2026-02-20 08:33:27.990538633 +0000 UTC m=+1656.965033463" observedRunningTime="2026-02-20 08:33:28.611033452 +0000 UTC m=+1657.585528272" watchObservedRunningTime="2026-02-20 08:33:28.612310293 +0000 UTC m=+1657.586805123" Feb 20 08:33:33 crc kubenswrapper[4948]: I0220 08:33:33.895660 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:33 crc kubenswrapper[4948]: I0220 08:33:33.896338 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:33 crc kubenswrapper[4948]: I0220 08:33:33.980385 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:34 crc kubenswrapper[4948]: I0220 08:33:34.744601 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:34 crc kubenswrapper[4948]: I0220 08:33:34.820609 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lvjzf"] Feb 20 08:33:36 crc kubenswrapper[4948]: I0220 08:33:36.674646 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lvjzf" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerName="registry-server" containerID="cri-o://2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d" gracePeriod=2 Feb 20 08:33:36 crc kubenswrapper[4948]: I0220 08:33:36.722776 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:33:36 crc kubenswrapper[4948]: E0220 08:33:36.723047 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.191211 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.285607 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-utilities\") pod \"26bee9f7-36e7-4452-be46-a7af14fe0d82\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.285659 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jfnl\" (UniqueName: \"kubernetes.io/projected/26bee9f7-36e7-4452-be46-a7af14fe0d82-kube-api-access-2jfnl\") pod \"26bee9f7-36e7-4452-be46-a7af14fe0d82\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.285787 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-catalog-content\") pod \"26bee9f7-36e7-4452-be46-a7af14fe0d82\" (UID: \"26bee9f7-36e7-4452-be46-a7af14fe0d82\") " Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.286748 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-utilities" (OuterVolumeSpecName: "utilities") pod "26bee9f7-36e7-4452-be46-a7af14fe0d82" (UID: "26bee9f7-36e7-4452-be46-a7af14fe0d82"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.292173 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26bee9f7-36e7-4452-be46-a7af14fe0d82-kube-api-access-2jfnl" (OuterVolumeSpecName: "kube-api-access-2jfnl") pod "26bee9f7-36e7-4452-be46-a7af14fe0d82" (UID: "26bee9f7-36e7-4452-be46-a7af14fe0d82"). InnerVolumeSpecName "kube-api-access-2jfnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.335618 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "26bee9f7-36e7-4452-be46-a7af14fe0d82" (UID: "26bee9f7-36e7-4452-be46-a7af14fe0d82"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.388236 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.388293 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jfnl\" (UniqueName: \"kubernetes.io/projected/26bee9f7-36e7-4452-be46-a7af14fe0d82-kube-api-access-2jfnl\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.388316 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26bee9f7-36e7-4452-be46-a7af14fe0d82-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.684757 4948 generic.go:334] "Generic (PLEG): container finished" podID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerID="2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d" exitCode=0 Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.684843 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lvjzf" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.684866 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvjzf" event={"ID":"26bee9f7-36e7-4452-be46-a7af14fe0d82","Type":"ContainerDied","Data":"2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d"} Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.685214 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvjzf" event={"ID":"26bee9f7-36e7-4452-be46-a7af14fe0d82","Type":"ContainerDied","Data":"d87e55975b4c132e628eaad19e343e2dc1a36778a825ecd32677e5386de184a8"} Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.685282 4948 scope.go:117] "RemoveContainer" containerID="2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.721260 4948 scope.go:117] "RemoveContainer" containerID="3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.747899 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lvjzf"] Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.756417 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lvjzf"] Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.758509 4948 scope.go:117] "RemoveContainer" containerID="d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.818181 4948 scope.go:117] "RemoveContainer" containerID="2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d" Feb 20 08:33:37 crc kubenswrapper[4948]: E0220 08:33:37.818738 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d\": container with ID starting with 2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d not found: ID does not exist" containerID="2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.818776 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d"} err="failed to get container status \"2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d\": rpc error: code = NotFound desc = could not find container \"2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d\": container with ID starting with 2ed5eb1d692e2406cfd129f46d82525b4996228bb8988814eef3b67af71f3d1d not found: ID does not exist" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.818799 4948 scope.go:117] "RemoveContainer" containerID="3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c" Feb 20 08:33:37 crc kubenswrapper[4948]: E0220 08:33:37.819071 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c\": container with ID starting with 3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c not found: ID does not exist" containerID="3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.819099 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c"} err="failed to get container status \"3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c\": rpc error: code = NotFound desc = could not find container \"3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c\": container with ID starting with 3bd508455cdc61cc2aed3667a9786384d7d8905d6c881e37112bbb71d52d5a9c not found: ID does not exist" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.819118 4948 scope.go:117] "RemoveContainer" containerID="d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a" Feb 20 08:33:37 crc kubenswrapper[4948]: E0220 08:33:37.819368 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a\": container with ID starting with d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a not found: ID does not exist" containerID="d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a" Feb 20 08:33:37 crc kubenswrapper[4948]: I0220 08:33:37.819390 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a"} err="failed to get container status \"d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a\": rpc error: code = NotFound desc = could not find container \"d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a\": container with ID starting with d1d54224a90227b551f9ae7bab7701dda7f4f8e05b155c101b1a90b05f80f77a not found: ID does not exist" Feb 20 08:33:38 crc kubenswrapper[4948]: I0220 08:33:38.085540 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-aecf-account-create-update-gnnvg"] Feb 20 08:33:38 crc kubenswrapper[4948]: I0220 08:33:38.102603 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-mgnlb"] Feb 20 08:33:38 crc kubenswrapper[4948]: I0220 08:33:38.111189 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-2rgc6"] Feb 20 08:33:38 crc kubenswrapper[4948]: I0220 08:33:38.120715 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-ca33-account-create-update-dwqpd"] Feb 20 08:33:38 crc kubenswrapper[4948]: I0220 08:33:38.130573 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-aecf-account-create-update-gnnvg"] Feb 20 08:33:38 crc kubenswrapper[4948]: I0220 08:33:38.137149 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-2rgc6"] Feb 20 08:33:38 crc kubenswrapper[4948]: I0220 08:33:38.143398 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-mgnlb"] Feb 20 08:33:38 crc kubenswrapper[4948]: I0220 08:33:38.149361 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-ca33-account-create-update-dwqpd"] Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.045603 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-x2pm7"] Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.059173 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-78de-account-create-update-fwjdc"] Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.072535 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-x2pm7"] Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.082966 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-78de-account-create-update-fwjdc"] Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.737454 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b61242-b8c1-4e1b-b319-3eafb79ef048" path="/var/lib/kubelet/pods/20b61242-b8c1-4e1b-b319-3eafb79ef048/volumes" Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.738671 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" path="/var/lib/kubelet/pods/26bee9f7-36e7-4452-be46-a7af14fe0d82/volumes" Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.740372 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a311ded-ab56-4f07-8b7f-9dcc4d70f647" path="/var/lib/kubelet/pods/8a311ded-ab56-4f07-8b7f-9dcc4d70f647/volumes" Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.742429 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b167b56-655a-4220-9823-8cc606f5f034" path="/var/lib/kubelet/pods/8b167b56-655a-4220-9823-8cc606f5f034/volumes" Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.743555 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5cb8c15-68fc-47db-9f68-1a8401f99f88" path="/var/lib/kubelet/pods/a5cb8c15-68fc-47db-9f68-1a8401f99f88/volumes" Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.744677 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f" path="/var/lib/kubelet/pods/b9f64dd5-5fb8-43e2-9c52-fc6d37ffec6f/volumes" Feb 20 08:33:39 crc kubenswrapper[4948]: I0220 08:33:39.745776 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa398740-9a54-472f-9736-0959dd54b657" path="/var/lib/kubelet/pods/fa398740-9a54-472f-9736-0959dd54b657/volumes" Feb 20 08:33:48 crc kubenswrapper[4948]: I0220 08:33:48.723047 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:33:48 crc kubenswrapper[4948]: E0220 08:33:48.724192 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:33:48 crc kubenswrapper[4948]: I0220 08:33:48.812086 4948 generic.go:334] "Generic (PLEG): container finished" podID="f42ddd0c-a2ec-414a-812d-f87213a3226d" containerID="ddcb74ec6513e9050b354f5584993e70b4784d82329c695b8cbcea156ddada02" exitCode=0 Feb 20 08:33:48 crc kubenswrapper[4948]: I0220 08:33:48.812149 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" event={"ID":"f42ddd0c-a2ec-414a-812d-f87213a3226d","Type":"ContainerDied","Data":"ddcb74ec6513e9050b354f5584993e70b4784d82329c695b8cbcea156ddada02"} Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.354065 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.472490 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-inventory\") pod \"f42ddd0c-a2ec-414a-812d-f87213a3226d\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.472594 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-ssh-key-openstack-edpm-ipam\") pod \"f42ddd0c-a2ec-414a-812d-f87213a3226d\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.472662 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxqrc\" (UniqueName: \"kubernetes.io/projected/f42ddd0c-a2ec-414a-812d-f87213a3226d-kube-api-access-dxqrc\") pod \"f42ddd0c-a2ec-414a-812d-f87213a3226d\" (UID: \"f42ddd0c-a2ec-414a-812d-f87213a3226d\") " Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.478725 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f42ddd0c-a2ec-414a-812d-f87213a3226d-kube-api-access-dxqrc" (OuterVolumeSpecName: "kube-api-access-dxqrc") pod "f42ddd0c-a2ec-414a-812d-f87213a3226d" (UID: "f42ddd0c-a2ec-414a-812d-f87213a3226d"). InnerVolumeSpecName "kube-api-access-dxqrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.520184 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f42ddd0c-a2ec-414a-812d-f87213a3226d" (UID: "f42ddd0c-a2ec-414a-812d-f87213a3226d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.531499 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-inventory" (OuterVolumeSpecName: "inventory") pod "f42ddd0c-a2ec-414a-812d-f87213a3226d" (UID: "f42ddd0c-a2ec-414a-812d-f87213a3226d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.575113 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.575159 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f42ddd0c-a2ec-414a-812d-f87213a3226d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.575178 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxqrc\" (UniqueName: \"kubernetes.io/projected/f42ddd0c-a2ec-414a-812d-f87213a3226d-kube-api-access-dxqrc\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.835800 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" event={"ID":"f42ddd0c-a2ec-414a-812d-f87213a3226d","Type":"ContainerDied","Data":"26e225a9996f58185c76eebc01d4b0d42e12e26c4c12d0fd5527a74efcfa0e7f"} Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.835861 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26e225a9996f58185c76eebc01d4b0d42e12e26c4c12d0fd5527a74efcfa0e7f" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.835937 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l55v" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.998821 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6"] Feb 20 08:33:50 crc kubenswrapper[4948]: E0220 08:33:50.999806 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerName="extract-utilities" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.999836 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerName="extract-utilities" Feb 20 08:33:50 crc kubenswrapper[4948]: E0220 08:33:50.999883 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerName="registry-server" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.999896 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerName="registry-server" Feb 20 08:33:50 crc kubenswrapper[4948]: E0220 08:33:50.999922 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerName="extract-content" Feb 20 08:33:50 crc kubenswrapper[4948]: I0220 08:33:50.999935 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerName="extract-content" Feb 20 08:33:51 crc kubenswrapper[4948]: E0220 08:33:50.999961 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f42ddd0c-a2ec-414a-812d-f87213a3226d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.000008 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f42ddd0c-a2ec-414a-812d-f87213a3226d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.000431 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="26bee9f7-36e7-4452-be46-a7af14fe0d82" containerName="registry-server" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.000472 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f42ddd0c-a2ec-414a-812d-f87213a3226d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.001373 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.003543 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.007463 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.007536 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.007687 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.026551 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6"] Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.186667 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8gm4\" (UniqueName: \"kubernetes.io/projected/d38195a4-0c7b-427e-b5e2-923b24c10674-kube-api-access-p8gm4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.186915 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.187129 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.289098 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.289182 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.289426 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8gm4\" (UniqueName: \"kubernetes.io/projected/d38195a4-0c7b-427e-b5e2-923b24c10674-kube-api-access-p8gm4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.295810 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.301085 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.325307 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8gm4\" (UniqueName: \"kubernetes.io/projected/d38195a4-0c7b-427e-b5e2-923b24c10674-kube-api-access-p8gm4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:51 crc kubenswrapper[4948]: I0220 08:33:51.621320 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:33:52 crc kubenswrapper[4948]: I0220 08:33:52.064255 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6"] Feb 20 08:33:52 crc kubenswrapper[4948]: I0220 08:33:52.473833 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:33:52 crc kubenswrapper[4948]: I0220 08:33:52.864779 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" event={"ID":"d38195a4-0c7b-427e-b5e2-923b24c10674","Type":"ContainerStarted","Data":"22f28812074b1280b97152f9c9a34b69bc5023788861680ff29864ecb44039df"} Feb 20 08:33:52 crc kubenswrapper[4948]: I0220 08:33:52.865354 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" event={"ID":"d38195a4-0c7b-427e-b5e2-923b24c10674","Type":"ContainerStarted","Data":"f1bf8524d66477c9c879feaa8dab8a73e46ed694dc686bb9ceaf259ba30a6e1e"} Feb 20 08:33:52 crc kubenswrapper[4948]: I0220 08:33:52.897896 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" podStartSLOduration=2.496455098 podStartE2EDuration="2.897878807s" podCreationTimestamp="2026-02-20 08:33:50 +0000 UTC" firstStartedPulling="2026-02-20 08:33:52.070076961 +0000 UTC m=+1681.044571781" lastFinishedPulling="2026-02-20 08:33:52.47150067 +0000 UTC m=+1681.445995490" observedRunningTime="2026-02-20 08:33:52.888459683 +0000 UTC m=+1681.862954503" watchObservedRunningTime="2026-02-20 08:33:52.897878807 +0000 UTC m=+1681.872373627" Feb 20 08:34:00 crc kubenswrapper[4948]: I0220 08:34:00.759316 4948 scope.go:117] "RemoveContainer" containerID="97e6bc167799c61927aab7b946388571d89bdf96214e87d2ba307142c0a90dd5" Feb 20 08:34:00 crc kubenswrapper[4948]: I0220 08:34:00.807342 4948 scope.go:117] "RemoveContainer" containerID="57a582df620782031dd517ee8549cae414c4157e56c0fc270434b874a8cbfef6" Feb 20 08:34:00 crc kubenswrapper[4948]: I0220 08:34:00.862758 4948 scope.go:117] "RemoveContainer" containerID="75747717ce5381d273c3b7d9687d0d98efbbf16a7bbea02c0a10888c6dd3a064" Feb 20 08:34:00 crc kubenswrapper[4948]: I0220 08:34:00.906947 4948 scope.go:117] "RemoveContainer" containerID="be6725bc5e2da748cc92ca6d3d54cea0585e5ca878e37159db354dc55a1eae54" Feb 20 08:34:00 crc kubenswrapper[4948]: I0220 08:34:00.955518 4948 scope.go:117] "RemoveContainer" containerID="dd9bca0f8ae0d62df16de42592e7d1d0793a43cd05a5a5d5dcaf73199e707005" Feb 20 08:34:00 crc kubenswrapper[4948]: I0220 08:34:00.997426 4948 scope.go:117] "RemoveContainer" containerID="b6240fac3873ea6c61937ec67473ce610d9e47448d45e4e229e635b6cb80cfa1" Feb 20 08:34:01 crc kubenswrapper[4948]: I0220 08:34:01.078735 4948 scope.go:117] "RemoveContainer" containerID="ca086fdde2d4c69122e9c700d4d4af3b8218c685639bf6df2d417cc08c1e51d3" Feb 20 08:34:01 crc kubenswrapper[4948]: I0220 08:34:01.112389 4948 scope.go:117] "RemoveContainer" containerID="5b0b3d5fbb40595687dd65feb8573194de8a92b25615987166c82509c29136f4" Feb 20 08:34:01 crc kubenswrapper[4948]: I0220 08:34:01.733883 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:34:01 crc kubenswrapper[4948]: E0220 08:34:01.734968 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:34:08 crc kubenswrapper[4948]: I0220 08:34:08.059163 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xt5t7"] Feb 20 08:34:08 crc kubenswrapper[4948]: I0220 08:34:08.075956 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xt5t7"] Feb 20 08:34:09 crc kubenswrapper[4948]: I0220 08:34:09.741384 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fdb2e02-8713-4205-ae31-0af33f8e4047" path="/var/lib/kubelet/pods/3fdb2e02-8713-4205-ae31-0af33f8e4047/volumes" Feb 20 08:34:12 crc kubenswrapper[4948]: I0220 08:34:12.722282 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:34:12 crc kubenswrapper[4948]: E0220 08:34:12.723339 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:34:25 crc kubenswrapper[4948]: I0220 08:34:25.723025 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:34:25 crc kubenswrapper[4948]: E0220 08:34:25.724087 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:34:29 crc kubenswrapper[4948]: I0220 08:34:29.049915 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-qnlrn"] Feb 20 08:34:29 crc kubenswrapper[4948]: I0220 08:34:29.069709 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-qnlrn"] Feb 20 08:34:29 crc kubenswrapper[4948]: I0220 08:34:29.741889 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b" path="/var/lib/kubelet/pods/21a33b1f-54ff-46d3-9ff1-bbe8398d8c2b/volumes" Feb 20 08:34:30 crc kubenswrapper[4948]: I0220 08:34:30.038088 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jl8ck"] Feb 20 08:34:30 crc kubenswrapper[4948]: I0220 08:34:30.056727 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jl8ck"] Feb 20 08:34:31 crc kubenswrapper[4948]: I0220 08:34:31.736682 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93f8e582-bd6a-44d4-a203-ba9950efcada" path="/var/lib/kubelet/pods/93f8e582-bd6a-44d4-a203-ba9950efcada/volumes" Feb 20 08:34:38 crc kubenswrapper[4948]: I0220 08:34:38.722914 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:34:38 crc kubenswrapper[4948]: E0220 08:34:38.724276 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:34:43 crc kubenswrapper[4948]: I0220 08:34:43.452781 4948 generic.go:334] "Generic (PLEG): container finished" podID="d38195a4-0c7b-427e-b5e2-923b24c10674" containerID="22f28812074b1280b97152f9c9a34b69bc5023788861680ff29864ecb44039df" exitCode=0 Feb 20 08:34:43 crc kubenswrapper[4948]: I0220 08:34:43.452890 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" event={"ID":"d38195a4-0c7b-427e-b5e2-923b24c10674","Type":"ContainerDied","Data":"22f28812074b1280b97152f9c9a34b69bc5023788861680ff29864ecb44039df"} Feb 20 08:34:44 crc kubenswrapper[4948]: I0220 08:34:44.976495 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.036358 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-ssh-key-openstack-edpm-ipam\") pod \"d38195a4-0c7b-427e-b5e2-923b24c10674\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.036417 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8gm4\" (UniqueName: \"kubernetes.io/projected/d38195a4-0c7b-427e-b5e2-923b24c10674-kube-api-access-p8gm4\") pod \"d38195a4-0c7b-427e-b5e2-923b24c10674\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.036696 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-inventory\") pod \"d38195a4-0c7b-427e-b5e2-923b24c10674\" (UID: \"d38195a4-0c7b-427e-b5e2-923b24c10674\") " Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.042616 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d38195a4-0c7b-427e-b5e2-923b24c10674-kube-api-access-p8gm4" (OuterVolumeSpecName: "kube-api-access-p8gm4") pod "d38195a4-0c7b-427e-b5e2-923b24c10674" (UID: "d38195a4-0c7b-427e-b5e2-923b24c10674"). InnerVolumeSpecName "kube-api-access-p8gm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.066162 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-inventory" (OuterVolumeSpecName: "inventory") pod "d38195a4-0c7b-427e-b5e2-923b24c10674" (UID: "d38195a4-0c7b-427e-b5e2-923b24c10674"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.071773 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d38195a4-0c7b-427e-b5e2-923b24c10674" (UID: "d38195a4-0c7b-427e-b5e2-923b24c10674"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.138962 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.139011 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d38195a4-0c7b-427e-b5e2-923b24c10674-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.139023 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8gm4\" (UniqueName: \"kubernetes.io/projected/d38195a4-0c7b-427e-b5e2-923b24c10674-kube-api-access-p8gm4\") on node \"crc\" DevicePath \"\"" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.480622 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" event={"ID":"d38195a4-0c7b-427e-b5e2-923b24c10674","Type":"ContainerDied","Data":"f1bf8524d66477c9c879feaa8dab8a73e46ed694dc686bb9ceaf259ba30a6e1e"} Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.480676 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1bf8524d66477c9c879feaa8dab8a73e46ed694dc686bb9ceaf259ba30a6e1e" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.480699 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.592670 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cqchc"] Feb 20 08:34:45 crc kubenswrapper[4948]: E0220 08:34:45.593116 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d38195a4-0c7b-427e-b5e2-923b24c10674" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.593135 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d38195a4-0c7b-427e-b5e2-923b24c10674" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.593382 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d38195a4-0c7b-427e-b5e2-923b24c10674" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.594157 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.596906 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.597350 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.597457 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.599415 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.603648 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cqchc"] Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.648592 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.648872 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.649086 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsvbm\" (UniqueName: \"kubernetes.io/projected/41856163-b621-439a-95c1-ca8d13f0a08c-kube-api-access-fsvbm\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.751327 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsvbm\" (UniqueName: \"kubernetes.io/projected/41856163-b621-439a-95c1-ca8d13f0a08c-kube-api-access-fsvbm\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.753857 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.754404 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.757235 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.758723 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.774742 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsvbm\" (UniqueName: \"kubernetes.io/projected/41856163-b621-439a-95c1-ca8d13f0a08c-kube-api-access-fsvbm\") pod \"ssh-known-hosts-edpm-deployment-cqchc\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:45 crc kubenswrapper[4948]: I0220 08:34:45.914432 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:46 crc kubenswrapper[4948]: I0220 08:34:46.439077 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cqchc"] Feb 20 08:34:46 crc kubenswrapper[4948]: I0220 08:34:46.494705 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" event={"ID":"41856163-b621-439a-95c1-ca8d13f0a08c","Type":"ContainerStarted","Data":"f7c8ce188843d7fd25ac7b234d3e54139f992328b052544c97a2a433dce34ce0"} Feb 20 08:34:47 crc kubenswrapper[4948]: I0220 08:34:47.507370 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" event={"ID":"41856163-b621-439a-95c1-ca8d13f0a08c","Type":"ContainerStarted","Data":"180dc0cc05cc1856f0bc16e7267d6cbb4c78efff33f86f297dc54c1848d855ae"} Feb 20 08:34:47 crc kubenswrapper[4948]: I0220 08:34:47.533497 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" podStartSLOduration=2.004716972 podStartE2EDuration="2.533475917s" podCreationTimestamp="2026-02-20 08:34:45 +0000 UTC" firstStartedPulling="2026-02-20 08:34:46.445201645 +0000 UTC m=+1735.419696485" lastFinishedPulling="2026-02-20 08:34:46.9739606 +0000 UTC m=+1735.948455430" observedRunningTime="2026-02-20 08:34:47.529871778 +0000 UTC m=+1736.504366618" watchObservedRunningTime="2026-02-20 08:34:47.533475917 +0000 UTC m=+1736.507970747" Feb 20 08:34:49 crc kubenswrapper[4948]: I0220 08:34:49.723580 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:34:49 crc kubenswrapper[4948]: E0220 08:34:49.724346 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:34:54 crc kubenswrapper[4948]: I0220 08:34:54.574321 4948 generic.go:334] "Generic (PLEG): container finished" podID="41856163-b621-439a-95c1-ca8d13f0a08c" containerID="180dc0cc05cc1856f0bc16e7267d6cbb4c78efff33f86f297dc54c1848d855ae" exitCode=0 Feb 20 08:34:54 crc kubenswrapper[4948]: I0220 08:34:54.574741 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" event={"ID":"41856163-b621-439a-95c1-ca8d13f0a08c","Type":"ContainerDied","Data":"180dc0cc05cc1856f0bc16e7267d6cbb4c78efff33f86f297dc54c1848d855ae"} Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.039710 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.173203 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-inventory-0\") pod \"41856163-b621-439a-95c1-ca8d13f0a08c\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.173355 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-ssh-key-openstack-edpm-ipam\") pod \"41856163-b621-439a-95c1-ca8d13f0a08c\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.173505 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsvbm\" (UniqueName: \"kubernetes.io/projected/41856163-b621-439a-95c1-ca8d13f0a08c-kube-api-access-fsvbm\") pod \"41856163-b621-439a-95c1-ca8d13f0a08c\" (UID: \"41856163-b621-439a-95c1-ca8d13f0a08c\") " Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.178720 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41856163-b621-439a-95c1-ca8d13f0a08c-kube-api-access-fsvbm" (OuterVolumeSpecName: "kube-api-access-fsvbm") pod "41856163-b621-439a-95c1-ca8d13f0a08c" (UID: "41856163-b621-439a-95c1-ca8d13f0a08c"). InnerVolumeSpecName "kube-api-access-fsvbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.200767 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "41856163-b621-439a-95c1-ca8d13f0a08c" (UID: "41856163-b621-439a-95c1-ca8d13f0a08c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.213865 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "41856163-b621-439a-95c1-ca8d13f0a08c" (UID: "41856163-b621-439a-95c1-ca8d13f0a08c"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.275283 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsvbm\" (UniqueName: \"kubernetes.io/projected/41856163-b621-439a-95c1-ca8d13f0a08c-kube-api-access-fsvbm\") on node \"crc\" DevicePath \"\"" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.275317 4948 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-inventory-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.275326 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41856163-b621-439a-95c1-ca8d13f0a08c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.590736 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" event={"ID":"41856163-b621-439a-95c1-ca8d13f0a08c","Type":"ContainerDied","Data":"f7c8ce188843d7fd25ac7b234d3e54139f992328b052544c97a2a433dce34ce0"} Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.590779 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7c8ce188843d7fd25ac7b234d3e54139f992328b052544c97a2a433dce34ce0" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.590808 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cqchc" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.683519 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf"] Feb 20 08:34:56 crc kubenswrapper[4948]: E0220 08:34:56.683952 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41856163-b621-439a-95c1-ca8d13f0a08c" containerName="ssh-known-hosts-edpm-deployment" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.684091 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="41856163-b621-439a-95c1-ca8d13f0a08c" containerName="ssh-known-hosts-edpm-deployment" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.684369 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="41856163-b621-439a-95c1-ca8d13f0a08c" containerName="ssh-known-hosts-edpm-deployment" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.685106 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.688753 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.690184 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.690732 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.690736 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.705110 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf"] Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.787548 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.787749 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.787850 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lxdt\" (UniqueName: \"kubernetes.io/projected/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-kube-api-access-7lxdt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.890533 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.890897 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.891187 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lxdt\" (UniqueName: \"kubernetes.io/projected/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-kube-api-access-7lxdt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.900496 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.903261 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:56 crc kubenswrapper[4948]: I0220 08:34:56.933063 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lxdt\" (UniqueName: \"kubernetes.io/projected/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-kube-api-access-7lxdt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lkbkf\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:57 crc kubenswrapper[4948]: I0220 08:34:57.004472 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:34:57 crc kubenswrapper[4948]: I0220 08:34:57.551737 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf"] Feb 20 08:34:57 crc kubenswrapper[4948]: I0220 08:34:57.600835 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" event={"ID":"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8","Type":"ContainerStarted","Data":"d5a7e5b1b8b0989f19fb68130988dab642729d856e78aedf79aafad894716612"} Feb 20 08:34:58 crc kubenswrapper[4948]: I0220 08:34:58.612221 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" event={"ID":"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8","Type":"ContainerStarted","Data":"40f85bc69eb4ee43a735e4ca0fe70430eaa230f800b74744b7b63478abaf2dd8"} Feb 20 08:34:58 crc kubenswrapper[4948]: I0220 08:34:58.635880 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" podStartSLOduration=2.1843637 podStartE2EDuration="2.635859696s" podCreationTimestamp="2026-02-20 08:34:56 +0000 UTC" firstStartedPulling="2026-02-20 08:34:57.546679039 +0000 UTC m=+1746.521173859" lastFinishedPulling="2026-02-20 08:34:57.998175025 +0000 UTC m=+1746.972669855" observedRunningTime="2026-02-20 08:34:58.633739683 +0000 UTC m=+1747.608234503" watchObservedRunningTime="2026-02-20 08:34:58.635859696 +0000 UTC m=+1747.610354516" Feb 20 08:35:01 crc kubenswrapper[4948]: I0220 08:35:01.399233 4948 scope.go:117] "RemoveContainer" containerID="9a9d495cbe6ddcf0246661f3edacc2b8afe21e8ef43eca82e9140a67f21f1c6d" Feb 20 08:35:01 crc kubenswrapper[4948]: I0220 08:35:01.468576 4948 scope.go:117] "RemoveContainer" containerID="67cb9d3659cf07e0efb010d75eb62a0dddad421b940140c438f0a0714676a1ed" Feb 20 08:35:01 crc kubenswrapper[4948]: I0220 08:35:01.514898 4948 scope.go:117] "RemoveContainer" containerID="53a1cca1717eaf7448d7d84b4b68c8bad0f23b2d1cc07d696f16d15b417b8356" Feb 20 08:35:01 crc kubenswrapper[4948]: I0220 08:35:01.730109 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:35:01 crc kubenswrapper[4948]: E0220 08:35:01.730440 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:35:06 crc kubenswrapper[4948]: I0220 08:35:06.696871 4948 generic.go:334] "Generic (PLEG): container finished" podID="79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8" containerID="40f85bc69eb4ee43a735e4ca0fe70430eaa230f800b74744b7b63478abaf2dd8" exitCode=0 Feb 20 08:35:06 crc kubenswrapper[4948]: I0220 08:35:06.697012 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" event={"ID":"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8","Type":"ContainerDied","Data":"40f85bc69eb4ee43a735e4ca0fe70430eaa230f800b74744b7b63478abaf2dd8"} Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.139403 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.222390 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-inventory\") pod \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.222463 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lxdt\" (UniqueName: \"kubernetes.io/projected/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-kube-api-access-7lxdt\") pod \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.222524 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam\") pod \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.227582 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-kube-api-access-7lxdt" (OuterVolumeSpecName: "kube-api-access-7lxdt") pod "79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8" (UID: "79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8"). InnerVolumeSpecName "kube-api-access-7lxdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:35:08 crc kubenswrapper[4948]: E0220 08:35:08.252320 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam podName:79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8 nodeName:}" failed. No retries permitted until 2026-02-20 08:35:08.752290018 +0000 UTC m=+1757.726784828 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key-openstack-edpm-ipam" (UniqueName: "kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam") pod "79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8" (UID: "79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8") : error deleting /var/lib/kubelet/pods/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8/volume-subpaths: remove /var/lib/kubelet/pods/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8/volume-subpaths: no such file or directory Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.256022 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-inventory" (OuterVolumeSpecName: "inventory") pod "79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8" (UID: "79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.328262 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.328300 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lxdt\" (UniqueName: \"kubernetes.io/projected/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-kube-api-access-7lxdt\") on node \"crc\" DevicePath \"\"" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.727054 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" event={"ID":"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8","Type":"ContainerDied","Data":"d5a7e5b1b8b0989f19fb68130988dab642729d856e78aedf79aafad894716612"} Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.727120 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5a7e5b1b8b0989f19fb68130988dab642729d856e78aedf79aafad894716612" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.727078 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lkbkf" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.826943 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522"] Feb 20 08:35:08 crc kubenswrapper[4948]: E0220 08:35:08.827958 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.828037 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.828483 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.829889 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.837497 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522"] Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.841440 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam\") pod \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\" (UID: \"79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8\") " Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.847780 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8" (UID: "79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.944489 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.944833 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.945235 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjgmm\" (UniqueName: \"kubernetes.io/projected/e3b05342-f970-4fdb-9b1c-ff521e118f96-kube-api-access-pjgmm\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:08 crc kubenswrapper[4948]: I0220 08:35:08.945495 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:35:09 crc kubenswrapper[4948]: I0220 08:35:09.048244 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjgmm\" (UniqueName: \"kubernetes.io/projected/e3b05342-f970-4fdb-9b1c-ff521e118f96-kube-api-access-pjgmm\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:09 crc kubenswrapper[4948]: I0220 08:35:09.048557 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:09 crc kubenswrapper[4948]: I0220 08:35:09.048834 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:09 crc kubenswrapper[4948]: I0220 08:35:09.055555 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:09 crc kubenswrapper[4948]: I0220 08:35:09.056928 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:09 crc kubenswrapper[4948]: I0220 08:35:09.068825 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjgmm\" (UniqueName: \"kubernetes.io/projected/e3b05342-f970-4fdb-9b1c-ff521e118f96-kube-api-access-pjgmm\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-ds522\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:09 crc kubenswrapper[4948]: I0220 08:35:09.193990 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:10 crc kubenswrapper[4948]: I0220 08:35:09.998585 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522"] Feb 20 08:35:10 crc kubenswrapper[4948]: I0220 08:35:10.746270 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" event={"ID":"e3b05342-f970-4fdb-9b1c-ff521e118f96","Type":"ContainerStarted","Data":"0efd1de2d13cbcfa7bb9f687818d126c6dfe370e31b97b9d68feae9168d4d5a5"} Feb 20 08:35:10 crc kubenswrapper[4948]: I0220 08:35:10.746669 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" event={"ID":"e3b05342-f970-4fdb-9b1c-ff521e118f96","Type":"ContainerStarted","Data":"0af26b0e4338eeef5f6df928698713b0376131f10f3bb8116615033f43d3300b"} Feb 20 08:35:10 crc kubenswrapper[4948]: I0220 08:35:10.775748 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" podStartSLOduration=2.36727645 podStartE2EDuration="2.775727143s" podCreationTimestamp="2026-02-20 08:35:08 +0000 UTC" firstStartedPulling="2026-02-20 08:35:10.00318989 +0000 UTC m=+1758.977684710" lastFinishedPulling="2026-02-20 08:35:10.411640583 +0000 UTC m=+1759.386135403" observedRunningTime="2026-02-20 08:35:10.767126229 +0000 UTC m=+1759.741621049" watchObservedRunningTime="2026-02-20 08:35:10.775727143 +0000 UTC m=+1759.750221963" Feb 20 08:35:14 crc kubenswrapper[4948]: I0220 08:35:14.069050 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-mg24f"] Feb 20 08:35:14 crc kubenswrapper[4948]: I0220 08:35:14.085041 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-mg24f"] Feb 20 08:35:15 crc kubenswrapper[4948]: I0220 08:35:15.723508 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:35:15 crc kubenswrapper[4948]: E0220 08:35:15.724360 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:35:15 crc kubenswrapper[4948]: I0220 08:35:15.738295 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf549f46-7597-4399-addc-3bf3f5b63f99" path="/var/lib/kubelet/pods/cf549f46-7597-4399-addc-3bf3f5b63f99/volumes" Feb 20 08:35:20 crc kubenswrapper[4948]: I0220 08:35:20.855770 4948 generic.go:334] "Generic (PLEG): container finished" podID="e3b05342-f970-4fdb-9b1c-ff521e118f96" containerID="0efd1de2d13cbcfa7bb9f687818d126c6dfe370e31b97b9d68feae9168d4d5a5" exitCode=0 Feb 20 08:35:20 crc kubenswrapper[4948]: I0220 08:35:20.855913 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" event={"ID":"e3b05342-f970-4fdb-9b1c-ff521e118f96","Type":"ContainerDied","Data":"0efd1de2d13cbcfa7bb9f687818d126c6dfe370e31b97b9d68feae9168d4d5a5"} Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.356957 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.542762 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-inventory\") pod \"e3b05342-f970-4fdb-9b1c-ff521e118f96\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.542918 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-ssh-key-openstack-edpm-ipam\") pod \"e3b05342-f970-4fdb-9b1c-ff521e118f96\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.543034 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjgmm\" (UniqueName: \"kubernetes.io/projected/e3b05342-f970-4fdb-9b1c-ff521e118f96-kube-api-access-pjgmm\") pod \"e3b05342-f970-4fdb-9b1c-ff521e118f96\" (UID: \"e3b05342-f970-4fdb-9b1c-ff521e118f96\") " Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.552209 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3b05342-f970-4fdb-9b1c-ff521e118f96-kube-api-access-pjgmm" (OuterVolumeSpecName: "kube-api-access-pjgmm") pod "e3b05342-f970-4fdb-9b1c-ff521e118f96" (UID: "e3b05342-f970-4fdb-9b1c-ff521e118f96"). InnerVolumeSpecName "kube-api-access-pjgmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.583175 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e3b05342-f970-4fdb-9b1c-ff521e118f96" (UID: "e3b05342-f970-4fdb-9b1c-ff521e118f96"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.617134 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-inventory" (OuterVolumeSpecName: "inventory") pod "e3b05342-f970-4fdb-9b1c-ff521e118f96" (UID: "e3b05342-f970-4fdb-9b1c-ff521e118f96"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.648328 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.648362 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3b05342-f970-4fdb-9b1c-ff521e118f96-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.648372 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjgmm\" (UniqueName: \"kubernetes.io/projected/e3b05342-f970-4fdb-9b1c-ff521e118f96-kube-api-access-pjgmm\") on node \"crc\" DevicePath \"\"" Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.888012 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.891217 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-ds522" event={"ID":"e3b05342-f970-4fdb-9b1c-ff521e118f96","Type":"ContainerDied","Data":"0af26b0e4338eeef5f6df928698713b0376131f10f3bb8116615033f43d3300b"} Feb 20 08:35:22 crc kubenswrapper[4948]: I0220 08:35:22.891272 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0af26b0e4338eeef5f6df928698713b0376131f10f3bb8116615033f43d3300b" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.009182 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw"] Feb 20 08:35:23 crc kubenswrapper[4948]: E0220 08:35:23.009736 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b05342-f970-4fdb-9b1c-ff521e118f96" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.009762 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b05342-f970-4fdb-9b1c-ff521e118f96" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.010105 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3b05342-f970-4fdb-9b1c-ff521e118f96" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.011603 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.014121 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.014417 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.015056 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.015105 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.015227 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.015301 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.015600 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.015613 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.030742 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw"] Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158485 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158532 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158611 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158650 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158673 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158696 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158750 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158776 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lwbl\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-kube-api-access-4lwbl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158792 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.158814 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.159085 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.159181 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.159265 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.159320 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.262264 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.262424 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.262513 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.262559 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.262699 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.262754 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lwbl\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-kube-api-access-4lwbl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.262790 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.262863 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.263772 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.263841 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.263947 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.264005 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.264067 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.264114 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.268434 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.268587 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.269040 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.269220 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.269557 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.270087 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.270948 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.271598 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.272536 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.274251 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.281379 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.281639 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.282115 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.286244 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lwbl\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-kube-api-access-4lwbl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.337805 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:35:23 crc kubenswrapper[4948]: W0220 08:35:23.888786 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc517042a_5354_4a8d_b7ff_c9aafa263b6c.slice/crio-835aa7fbf4e7707cff80ac88d34e7f17f33d8271cc0cd95b2c1162f2db4176aa WatchSource:0}: Error finding container 835aa7fbf4e7707cff80ac88d34e7f17f33d8271cc0cd95b2c1162f2db4176aa: Status 404 returned error can't find the container with id 835aa7fbf4e7707cff80ac88d34e7f17f33d8271cc0cd95b2c1162f2db4176aa Feb 20 08:35:23 crc kubenswrapper[4948]: I0220 08:35:23.904477 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw"] Feb 20 08:35:24 crc kubenswrapper[4948]: I0220 08:35:24.907421 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" event={"ID":"c517042a-5354-4a8d-b7ff-c9aafa263b6c","Type":"ContainerStarted","Data":"cc1c5248f2eeca932bf8566e96f4d010e0d2b251e1788d5c187c35a67338cca0"} Feb 20 08:35:24 crc kubenswrapper[4948]: I0220 08:35:24.907773 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" event={"ID":"c517042a-5354-4a8d-b7ff-c9aafa263b6c","Type":"ContainerStarted","Data":"835aa7fbf4e7707cff80ac88d34e7f17f33d8271cc0cd95b2c1162f2db4176aa"} Feb 20 08:35:24 crc kubenswrapper[4948]: I0220 08:35:24.928051 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" podStartSLOduration=2.531815314 podStartE2EDuration="2.928024752s" podCreationTimestamp="2026-02-20 08:35:22 +0000 UTC" firstStartedPulling="2026-02-20 08:35:23.892144914 +0000 UTC m=+1772.866639744" lastFinishedPulling="2026-02-20 08:35:24.288354322 +0000 UTC m=+1773.262849182" observedRunningTime="2026-02-20 08:35:24.92633649 +0000 UTC m=+1773.900831330" watchObservedRunningTime="2026-02-20 08:35:24.928024752 +0000 UTC m=+1773.902519612" Feb 20 08:35:26 crc kubenswrapper[4948]: I0220 08:35:26.723834 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:35:26 crc kubenswrapper[4948]: E0220 08:35:26.724594 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:35:41 crc kubenswrapper[4948]: I0220 08:35:41.732664 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:35:42 crc kubenswrapper[4948]: I0220 08:35:42.115016 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"df133c52e6c7c766068e5890945bf88c2349bad99b1086eddf879822786cef6f"} Feb 20 08:36:01 crc kubenswrapper[4948]: I0220 08:36:01.613602 4948 scope.go:117] "RemoveContainer" containerID="d68a0fb50da62e47bddf140b8e1628a7abf7e9ac68f7c4c4efaa6f9ccf62f179" Feb 20 08:36:02 crc kubenswrapper[4948]: I0220 08:36:02.318394 4948 generic.go:334] "Generic (PLEG): container finished" podID="c517042a-5354-4a8d-b7ff-c9aafa263b6c" containerID="cc1c5248f2eeca932bf8566e96f4d010e0d2b251e1788d5c187c35a67338cca0" exitCode=0 Feb 20 08:36:02 crc kubenswrapper[4948]: I0220 08:36:02.318532 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" event={"ID":"c517042a-5354-4a8d-b7ff-c9aafa263b6c","Type":"ContainerDied","Data":"cc1c5248f2eeca932bf8566e96f4d010e0d2b251e1788d5c187c35a67338cca0"} Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.880728 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.945020 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.945075 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-telemetry-combined-ca-bundle\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.945102 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-repo-setup-combined-ca-bundle\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.945126 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ovn-combined-ca-bundle\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.945159 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.945198 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-libvirt-combined-ca-bundle\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.945236 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-nova-combined-ca-bundle\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.945258 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.946363 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.946493 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-inventory\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.946534 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ssh-key-openstack-edpm-ipam\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.946571 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lwbl\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-kube-api-access-4lwbl\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.946599 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-bootstrap-combined-ca-bundle\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.946626 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-neutron-metadata-combined-ca-bundle\") pod \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\" (UID: \"c517042a-5354-4a8d-b7ff-c9aafa263b6c\") " Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.952675 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.953936 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.954147 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.954771 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.955617 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.955713 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.957586 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.957633 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.960061 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.960290 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.961217 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-kube-api-access-4lwbl" (OuterVolumeSpecName: "kube-api-access-4lwbl") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "kube-api-access-4lwbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.965323 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:03 crc kubenswrapper[4948]: I0220 08:36:03.994516 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-inventory" (OuterVolumeSpecName: "inventory") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.009925 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c517042a-5354-4a8d-b7ff-c9aafa263b6c" (UID: "c517042a-5354-4a8d-b7ff-c9aafa263b6c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.049021 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.049228 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.049349 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lwbl\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-kube-api-access-4lwbl\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.049487 4948 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.049600 4948 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.049715 4948 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.049900 4948 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.050107 4948 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.050248 4948 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.050376 4948 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.050492 4948 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.050604 4948 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c517042a-5354-4a8d-b7ff-c9aafa263b6c-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.050715 4948 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.050824 4948 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c517042a-5354-4a8d-b7ff-c9aafa263b6c-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.343515 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" event={"ID":"c517042a-5354-4a8d-b7ff-c9aafa263b6c","Type":"ContainerDied","Data":"835aa7fbf4e7707cff80ac88d34e7f17f33d8271cc0cd95b2c1162f2db4176aa"} Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.343933 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="835aa7fbf4e7707cff80ac88d34e7f17f33d8271cc0cd95b2c1162f2db4176aa" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.343577 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.468319 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp"] Feb 20 08:36:04 crc kubenswrapper[4948]: E0220 08:36:04.468785 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c517042a-5354-4a8d-b7ff-c9aafa263b6c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.468812 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="c517042a-5354-4a8d-b7ff-c9aafa263b6c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.469070 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="c517042a-5354-4a8d-b7ff-c9aafa263b6c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.469834 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.473298 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.473414 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.473304 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.474397 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.476194 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.485045 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp"] Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.561148 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.561209 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.561263 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mw76\" (UniqueName: \"kubernetes.io/projected/b19ca7cb-01d2-4965-90f6-5d50211870cc-kube-api-access-7mw76\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.561281 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.561324 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.662511 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.662599 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.662673 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mw76\" (UniqueName: \"kubernetes.io/projected/b19ca7cb-01d2-4965-90f6-5d50211870cc-kube-api-access-7mw76\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.662697 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.662754 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.664006 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.666577 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.667152 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.667615 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.684347 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mw76\" (UniqueName: \"kubernetes.io/projected/b19ca7cb-01d2-4965-90f6-5d50211870cc-kube-api-access-7mw76\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4b2vp\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:04 crc kubenswrapper[4948]: I0220 08:36:04.789493 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:36:05 crc kubenswrapper[4948]: I0220 08:36:05.353340 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp"] Feb 20 08:36:06 crc kubenswrapper[4948]: I0220 08:36:06.375141 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" event={"ID":"b19ca7cb-01d2-4965-90f6-5d50211870cc","Type":"ContainerStarted","Data":"3579159916fda6b8e313e027e891a641d962daeb27067a651a9d7a959ebf4296"} Feb 20 08:36:06 crc kubenswrapper[4948]: I0220 08:36:06.375715 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" event={"ID":"b19ca7cb-01d2-4965-90f6-5d50211870cc","Type":"ContainerStarted","Data":"03b19942c3e65950a34fa710ca9783f811998ce8fb63c17f12d02bef5e823c2b"} Feb 20 08:36:06 crc kubenswrapper[4948]: I0220 08:36:06.398863 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" podStartSLOduration=1.837710511 podStartE2EDuration="2.398845589s" podCreationTimestamp="2026-02-20 08:36:04 +0000 UTC" firstStartedPulling="2026-02-20 08:36:05.378967588 +0000 UTC m=+1814.353462408" lastFinishedPulling="2026-02-20 08:36:05.940102666 +0000 UTC m=+1814.914597486" observedRunningTime="2026-02-20 08:36:06.390092872 +0000 UTC m=+1815.364587732" watchObservedRunningTime="2026-02-20 08:36:06.398845589 +0000 UTC m=+1815.373340409" Feb 20 08:37:10 crc kubenswrapper[4948]: I0220 08:37:10.016466 4948 generic.go:334] "Generic (PLEG): container finished" podID="b19ca7cb-01d2-4965-90f6-5d50211870cc" containerID="3579159916fda6b8e313e027e891a641d962daeb27067a651a9d7a959ebf4296" exitCode=0 Feb 20 08:37:10 crc kubenswrapper[4948]: I0220 08:37:10.016577 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" event={"ID":"b19ca7cb-01d2-4965-90f6-5d50211870cc","Type":"ContainerDied","Data":"3579159916fda6b8e313e027e891a641d962daeb27067a651a9d7a959ebf4296"} Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.541597 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.717020 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-inventory\") pod \"b19ca7cb-01d2-4965-90f6-5d50211870cc\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.717324 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovncontroller-config-0\") pod \"b19ca7cb-01d2-4965-90f6-5d50211870cc\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.717388 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovn-combined-ca-bundle\") pod \"b19ca7cb-01d2-4965-90f6-5d50211870cc\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.717496 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mw76\" (UniqueName: \"kubernetes.io/projected/b19ca7cb-01d2-4965-90f6-5d50211870cc-kube-api-access-7mw76\") pod \"b19ca7cb-01d2-4965-90f6-5d50211870cc\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.717662 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ssh-key-openstack-edpm-ipam\") pod \"b19ca7cb-01d2-4965-90f6-5d50211870cc\" (UID: \"b19ca7cb-01d2-4965-90f6-5d50211870cc\") " Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.736120 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b19ca7cb-01d2-4965-90f6-5d50211870cc-kube-api-access-7mw76" (OuterVolumeSpecName: "kube-api-access-7mw76") pod "b19ca7cb-01d2-4965-90f6-5d50211870cc" (UID: "b19ca7cb-01d2-4965-90f6-5d50211870cc"). InnerVolumeSpecName "kube-api-access-7mw76". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.737501 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b19ca7cb-01d2-4965-90f6-5d50211870cc" (UID: "b19ca7cb-01d2-4965-90f6-5d50211870cc"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.747660 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-inventory" (OuterVolumeSpecName: "inventory") pod "b19ca7cb-01d2-4965-90f6-5d50211870cc" (UID: "b19ca7cb-01d2-4965-90f6-5d50211870cc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.764818 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b19ca7cb-01d2-4965-90f6-5d50211870cc" (UID: "b19ca7cb-01d2-4965-90f6-5d50211870cc"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.779251 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "b19ca7cb-01d2-4965-90f6-5d50211870cc" (UID: "b19ca7cb-01d2-4965-90f6-5d50211870cc"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.821009 4948 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.821080 4948 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.821109 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mw76\" (UniqueName: \"kubernetes.io/projected/b19ca7cb-01d2-4965-90f6-5d50211870cc-kube-api-access-7mw76\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.821137 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:11 crc kubenswrapper[4948]: I0220 08:37:11.821165 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b19ca7cb-01d2-4965-90f6-5d50211870cc-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.044536 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" event={"ID":"b19ca7cb-01d2-4965-90f6-5d50211870cc","Type":"ContainerDied","Data":"03b19942c3e65950a34fa710ca9783f811998ce8fb63c17f12d02bef5e823c2b"} Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.044598 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03b19942c3e65950a34fa710ca9783f811998ce8fb63c17f12d02bef5e823c2b" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.044662 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4b2vp" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.209884 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr"] Feb 20 08:37:12 crc kubenswrapper[4948]: E0220 08:37:12.210695 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b19ca7cb-01d2-4965-90f6-5d50211870cc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.210712 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b19ca7cb-01d2-4965-90f6-5d50211870cc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.210960 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b19ca7cb-01d2-4965-90f6-5d50211870cc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.211719 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.217011 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.217182 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.217280 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.217385 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.217494 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.217263 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.219864 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr"] Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.333156 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.333418 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.333583 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.333674 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.333783 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.334163 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp9df\" (UniqueName: \"kubernetes.io/projected/fd280342-4276-4eee-a763-876e31be28c5-kube-api-access-cp9df\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.435571 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.435824 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.435933 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.436028 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp9df\" (UniqueName: \"kubernetes.io/projected/fd280342-4276-4eee-a763-876e31be28c5-kube-api-access-cp9df\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.436124 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.436209 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.440773 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.441047 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.441467 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.442286 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.451481 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.454255 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp9df\" (UniqueName: \"kubernetes.io/projected/fd280342-4276-4eee-a763-876e31be28c5-kube-api-access-cp9df\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:12 crc kubenswrapper[4948]: I0220 08:37:12.539767 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:37:13 crc kubenswrapper[4948]: I0220 08:37:13.151203 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr"] Feb 20 08:37:14 crc kubenswrapper[4948]: I0220 08:37:14.070101 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" event={"ID":"fd280342-4276-4eee-a763-876e31be28c5","Type":"ContainerStarted","Data":"da36425e07d1062948867ebf1c1a2889de547fdeccbc610aebafa16fe2b84d8c"} Feb 20 08:37:14 crc kubenswrapper[4948]: I0220 08:37:14.070506 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" event={"ID":"fd280342-4276-4eee-a763-876e31be28c5","Type":"ContainerStarted","Data":"59f5aaf9278d5e594e4ceb19412a5a664c76f9cc8655925b2ee3a4a3ea09e8b3"} Feb 20 08:37:14 crc kubenswrapper[4948]: I0220 08:37:14.098397 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" podStartSLOduration=1.63944544 podStartE2EDuration="2.098372453s" podCreationTimestamp="2026-02-20 08:37:12 +0000 UTC" firstStartedPulling="2026-02-20 08:37:13.171435483 +0000 UTC m=+1882.145930313" lastFinishedPulling="2026-02-20 08:37:13.630362476 +0000 UTC m=+1882.604857326" observedRunningTime="2026-02-20 08:37:14.097283966 +0000 UTC m=+1883.071778846" watchObservedRunningTime="2026-02-20 08:37:14.098372453 +0000 UTC m=+1883.072867303" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.751678 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cqm4v"] Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.755783 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.766219 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cqm4v"] Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.849100 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-utilities\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.849189 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjrcl\" (UniqueName: \"kubernetes.io/projected/35e908da-6c2c-4e75-a567-7ad1b95dacf4-kube-api-access-hjrcl\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.849279 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-catalog-content\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.951464 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjrcl\" (UniqueName: \"kubernetes.io/projected/35e908da-6c2c-4e75-a567-7ad1b95dacf4-kube-api-access-hjrcl\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.951899 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-catalog-content\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.952190 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-utilities\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.952784 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-catalog-content\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.952812 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-utilities\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:30 crc kubenswrapper[4948]: I0220 08:37:30.982421 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjrcl\" (UniqueName: \"kubernetes.io/projected/35e908da-6c2c-4e75-a567-7ad1b95dacf4-kube-api-access-hjrcl\") pod \"redhat-marketplace-cqm4v\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:31 crc kubenswrapper[4948]: I0220 08:37:31.093032 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:31 crc kubenswrapper[4948]: I0220 08:37:31.645465 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cqm4v"] Feb 20 08:37:32 crc kubenswrapper[4948]: I0220 08:37:32.249110 4948 generic.go:334] "Generic (PLEG): container finished" podID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerID="cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8" exitCode=0 Feb 20 08:37:32 crc kubenswrapper[4948]: I0220 08:37:32.249227 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cqm4v" event={"ID":"35e908da-6c2c-4e75-a567-7ad1b95dacf4","Type":"ContainerDied","Data":"cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8"} Feb 20 08:37:32 crc kubenswrapper[4948]: I0220 08:37:32.249500 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cqm4v" event={"ID":"35e908da-6c2c-4e75-a567-7ad1b95dacf4","Type":"ContainerStarted","Data":"8bcf259501cac28be63c9a620abde433c308b13a601573640c689320a460160d"} Feb 20 08:37:33 crc kubenswrapper[4948]: I0220 08:37:33.268661 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cqm4v" event={"ID":"35e908da-6c2c-4e75-a567-7ad1b95dacf4","Type":"ContainerStarted","Data":"35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b"} Feb 20 08:37:34 crc kubenswrapper[4948]: I0220 08:37:34.284416 4948 generic.go:334] "Generic (PLEG): container finished" podID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerID="35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b" exitCode=0 Feb 20 08:37:34 crc kubenswrapper[4948]: I0220 08:37:34.284523 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cqm4v" event={"ID":"35e908da-6c2c-4e75-a567-7ad1b95dacf4","Type":"ContainerDied","Data":"35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b"} Feb 20 08:37:35 crc kubenswrapper[4948]: I0220 08:37:35.296597 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cqm4v" event={"ID":"35e908da-6c2c-4e75-a567-7ad1b95dacf4","Type":"ContainerStarted","Data":"b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952"} Feb 20 08:37:35 crc kubenswrapper[4948]: I0220 08:37:35.330640 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cqm4v" podStartSLOduration=2.8606195420000002 podStartE2EDuration="5.330620569s" podCreationTimestamp="2026-02-20 08:37:30 +0000 UTC" firstStartedPulling="2026-02-20 08:37:32.251540499 +0000 UTC m=+1901.226035399" lastFinishedPulling="2026-02-20 08:37:34.721541576 +0000 UTC m=+1903.696036426" observedRunningTime="2026-02-20 08:37:35.322468478 +0000 UTC m=+1904.296963318" watchObservedRunningTime="2026-02-20 08:37:35.330620569 +0000 UTC m=+1904.305115399" Feb 20 08:37:41 crc kubenswrapper[4948]: I0220 08:37:41.093245 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:41 crc kubenswrapper[4948]: I0220 08:37:41.094164 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:41 crc kubenswrapper[4948]: I0220 08:37:41.170861 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:41 crc kubenswrapper[4948]: I0220 08:37:41.422498 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:41 crc kubenswrapper[4948]: I0220 08:37:41.475644 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cqm4v"] Feb 20 08:37:43 crc kubenswrapper[4948]: I0220 08:37:43.375810 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cqm4v" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerName="registry-server" containerID="cri-o://b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952" gracePeriod=2 Feb 20 08:37:43 crc kubenswrapper[4948]: I0220 08:37:43.929272 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.091951 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-utilities\") pod \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.092109 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjrcl\" (UniqueName: \"kubernetes.io/projected/35e908da-6c2c-4e75-a567-7ad1b95dacf4-kube-api-access-hjrcl\") pod \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.092471 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-catalog-content\") pod \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\" (UID: \"35e908da-6c2c-4e75-a567-7ad1b95dacf4\") " Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.093428 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-utilities" (OuterVolumeSpecName: "utilities") pod "35e908da-6c2c-4e75-a567-7ad1b95dacf4" (UID: "35e908da-6c2c-4e75-a567-7ad1b95dacf4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.102472 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35e908da-6c2c-4e75-a567-7ad1b95dacf4-kube-api-access-hjrcl" (OuterVolumeSpecName: "kube-api-access-hjrcl") pod "35e908da-6c2c-4e75-a567-7ad1b95dacf4" (UID: "35e908da-6c2c-4e75-a567-7ad1b95dacf4"). InnerVolumeSpecName "kube-api-access-hjrcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.123166 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35e908da-6c2c-4e75-a567-7ad1b95dacf4" (UID: "35e908da-6c2c-4e75-a567-7ad1b95dacf4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.197191 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.197231 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35e908da-6c2c-4e75-a567-7ad1b95dacf4-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.197244 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjrcl\" (UniqueName: \"kubernetes.io/projected/35e908da-6c2c-4e75-a567-7ad1b95dacf4-kube-api-access-hjrcl\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.394851 4948 generic.go:334] "Generic (PLEG): container finished" podID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerID="b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952" exitCode=0 Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.394966 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cqm4v" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.395046 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cqm4v" event={"ID":"35e908da-6c2c-4e75-a567-7ad1b95dacf4","Type":"ContainerDied","Data":"b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952"} Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.396525 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cqm4v" event={"ID":"35e908da-6c2c-4e75-a567-7ad1b95dacf4","Type":"ContainerDied","Data":"8bcf259501cac28be63c9a620abde433c308b13a601573640c689320a460160d"} Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.396567 4948 scope.go:117] "RemoveContainer" containerID="b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.425878 4948 scope.go:117] "RemoveContainer" containerID="35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.460580 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cqm4v"] Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.471788 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cqm4v"] Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.472262 4948 scope.go:117] "RemoveContainer" containerID="cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.516323 4948 scope.go:117] "RemoveContainer" containerID="b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952" Feb 20 08:37:44 crc kubenswrapper[4948]: E0220 08:37:44.520741 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952\": container with ID starting with b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952 not found: ID does not exist" containerID="b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.520791 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952"} err="failed to get container status \"b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952\": rpc error: code = NotFound desc = could not find container \"b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952\": container with ID starting with b75e9f035613e5a0bca9527f0d1752421f2a649641c9c5f2a53156d6a7934952 not found: ID does not exist" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.520817 4948 scope.go:117] "RemoveContainer" containerID="35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b" Feb 20 08:37:44 crc kubenswrapper[4948]: E0220 08:37:44.521306 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b\": container with ID starting with 35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b not found: ID does not exist" containerID="35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.521355 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b"} err="failed to get container status \"35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b\": rpc error: code = NotFound desc = could not find container \"35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b\": container with ID starting with 35e1c15903ef72b4f5767d562be34dd2b8ca0c9c4040a73fcefd14e2c16aa84b not found: ID does not exist" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.521398 4948 scope.go:117] "RemoveContainer" containerID="cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8" Feb 20 08:37:44 crc kubenswrapper[4948]: E0220 08:37:44.521811 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8\": container with ID starting with cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8 not found: ID does not exist" containerID="cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8" Feb 20 08:37:44 crc kubenswrapper[4948]: I0220 08:37:44.521856 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8"} err="failed to get container status \"cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8\": rpc error: code = NotFound desc = could not find container \"cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8\": container with ID starting with cc4ee197b9a059416805589950a2af69d52f9384246abc136a26690a03566cc8 not found: ID does not exist" Feb 20 08:37:45 crc kubenswrapper[4948]: I0220 08:37:45.733993 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" path="/var/lib/kubelet/pods/35e908da-6c2c-4e75-a567-7ad1b95dacf4/volumes" Feb 20 08:38:02 crc kubenswrapper[4948]: I0220 08:38:02.596384 4948 generic.go:334] "Generic (PLEG): container finished" podID="fd280342-4276-4eee-a763-876e31be28c5" containerID="da36425e07d1062948867ebf1c1a2889de547fdeccbc610aebafa16fe2b84d8c" exitCode=0 Feb 20 08:38:02 crc kubenswrapper[4948]: I0220 08:38:02.596468 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" event={"ID":"fd280342-4276-4eee-a763-876e31be28c5","Type":"ContainerDied","Data":"da36425e07d1062948867ebf1c1a2889de547fdeccbc610aebafa16fe2b84d8c"} Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.210869 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.377784 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-inventory\") pod \"fd280342-4276-4eee-a763-876e31be28c5\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.378288 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-metadata-combined-ca-bundle\") pod \"fd280342-4276-4eee-a763-876e31be28c5\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.378339 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp9df\" (UniqueName: \"kubernetes.io/projected/fd280342-4276-4eee-a763-876e31be28c5-kube-api-access-cp9df\") pod \"fd280342-4276-4eee-a763-876e31be28c5\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.378566 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-ssh-key-openstack-edpm-ipam\") pod \"fd280342-4276-4eee-a763-876e31be28c5\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.378688 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-ovn-metadata-agent-neutron-config-0\") pod \"fd280342-4276-4eee-a763-876e31be28c5\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.378794 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-nova-metadata-neutron-config-0\") pod \"fd280342-4276-4eee-a763-876e31be28c5\" (UID: \"fd280342-4276-4eee-a763-876e31be28c5\") " Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.385747 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd280342-4276-4eee-a763-876e31be28c5-kube-api-access-cp9df" (OuterVolumeSpecName: "kube-api-access-cp9df") pod "fd280342-4276-4eee-a763-876e31be28c5" (UID: "fd280342-4276-4eee-a763-876e31be28c5"). InnerVolumeSpecName "kube-api-access-cp9df". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.385807 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "fd280342-4276-4eee-a763-876e31be28c5" (UID: "fd280342-4276-4eee-a763-876e31be28c5"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.418115 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-inventory" (OuterVolumeSpecName: "inventory") pod "fd280342-4276-4eee-a763-876e31be28c5" (UID: "fd280342-4276-4eee-a763-876e31be28c5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.419516 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "fd280342-4276-4eee-a763-876e31be28c5" (UID: "fd280342-4276-4eee-a763-876e31be28c5"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.429558 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "fd280342-4276-4eee-a763-876e31be28c5" (UID: "fd280342-4276-4eee-a763-876e31be28c5"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.441902 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "fd280342-4276-4eee-a763-876e31be28c5" (UID: "fd280342-4276-4eee-a763-876e31be28c5"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.482175 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.482215 4948 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.482276 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp9df\" (UniqueName: \"kubernetes.io/projected/fd280342-4276-4eee-a763-876e31be28c5-kube-api-access-cp9df\") on node \"crc\" DevicePath \"\"" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.482288 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.482323 4948 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.482337 4948 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fd280342-4276-4eee-a763-876e31be28c5-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.623260 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" event={"ID":"fd280342-4276-4eee-a763-876e31be28c5","Type":"ContainerDied","Data":"59f5aaf9278d5e594e4ceb19412a5a664c76f9cc8655925b2ee3a4a3ea09e8b3"} Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.623329 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59f5aaf9278d5e594e4ceb19412a5a664c76f9cc8655925b2ee3a4a3ea09e8b3" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.623334 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.727015 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs"] Feb 20 08:38:04 crc kubenswrapper[4948]: E0220 08:38:04.727695 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerName="extract-content" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.727788 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerName="extract-content" Feb 20 08:38:04 crc kubenswrapper[4948]: E0220 08:38:04.727876 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerName="registry-server" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.727954 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerName="registry-server" Feb 20 08:38:04 crc kubenswrapper[4948]: E0220 08:38:04.728068 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd280342-4276-4eee-a763-876e31be28c5" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.728140 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd280342-4276-4eee-a763-876e31be28c5" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 20 08:38:04 crc kubenswrapper[4948]: E0220 08:38:04.728222 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerName="extract-utilities" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.728289 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerName="extract-utilities" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.728607 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd280342-4276-4eee-a763-876e31be28c5" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.728706 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e908da-6c2c-4e75-a567-7ad1b95dacf4" containerName="registry-server" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.729491 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.732561 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.733237 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.734147 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.741834 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs"] Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.742295 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.745555 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.788714 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.788810 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.788940 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.789075 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8gnt\" (UniqueName: \"kubernetes.io/projected/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-kube-api-access-x8gnt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.789201 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.890186 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.890277 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8gnt\" (UniqueName: \"kubernetes.io/projected/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-kube-api-access-x8gnt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.890347 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.890609 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.890664 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.895904 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.896095 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.897248 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.897382 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:04 crc kubenswrapper[4948]: I0220 08:38:04.925366 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8gnt\" (UniqueName: \"kubernetes.io/projected/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-kube-api-access-x8gnt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:05 crc kubenswrapper[4948]: I0220 08:38:05.048938 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:38:05 crc kubenswrapper[4948]: I0220 08:38:05.595036 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs"] Feb 20 08:38:05 crc kubenswrapper[4948]: W0220 08:38:05.603183 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e619eac_dfdc_4c8c_aa51_fc08ba58b49a.slice/crio-bed584d1cffc5c44267aa8da79eb83c0547d67f904d60a696d0a5f0983625d4a WatchSource:0}: Error finding container bed584d1cffc5c44267aa8da79eb83c0547d67f904d60a696d0a5f0983625d4a: Status 404 returned error can't find the container with id bed584d1cffc5c44267aa8da79eb83c0547d67f904d60a696d0a5f0983625d4a Feb 20 08:38:05 crc kubenswrapper[4948]: I0220 08:38:05.607247 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:38:05 crc kubenswrapper[4948]: I0220 08:38:05.632956 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" event={"ID":"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a","Type":"ContainerStarted","Data":"bed584d1cffc5c44267aa8da79eb83c0547d67f904d60a696d0a5f0983625d4a"} Feb 20 08:38:06 crc kubenswrapper[4948]: I0220 08:38:06.648881 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" event={"ID":"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a","Type":"ContainerStarted","Data":"e26f1bc45952ca3a3414ed06ca781d747c45268c9f970dd0c34eb0f719572756"} Feb 20 08:38:06 crc kubenswrapper[4948]: I0220 08:38:06.688495 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" podStartSLOduration=2.252412711 podStartE2EDuration="2.688463518s" podCreationTimestamp="2026-02-20 08:38:04 +0000 UTC" firstStartedPulling="2026-02-20 08:38:05.606999149 +0000 UTC m=+1934.581493969" lastFinishedPulling="2026-02-20 08:38:06.043049946 +0000 UTC m=+1935.017544776" observedRunningTime="2026-02-20 08:38:06.675775225 +0000 UTC m=+1935.650270075" watchObservedRunningTime="2026-02-20 08:38:06.688463518 +0000 UTC m=+1935.662958378" Feb 20 08:38:08 crc kubenswrapper[4948]: I0220 08:38:08.025442 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:38:08 crc kubenswrapper[4948]: I0220 08:38:08.026627 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:38:38 crc kubenswrapper[4948]: I0220 08:38:38.025295 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:38:38 crc kubenswrapper[4948]: I0220 08:38:38.025837 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:39:08 crc kubenswrapper[4948]: I0220 08:39:08.024572 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:39:08 crc kubenswrapper[4948]: I0220 08:39:08.025402 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:39:08 crc kubenswrapper[4948]: I0220 08:39:08.025520 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:39:08 crc kubenswrapper[4948]: I0220 08:39:08.026750 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"df133c52e6c7c766068e5890945bf88c2349bad99b1086eddf879822786cef6f"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:39:08 crc kubenswrapper[4948]: I0220 08:39:08.026853 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://df133c52e6c7c766068e5890945bf88c2349bad99b1086eddf879822786cef6f" gracePeriod=600 Feb 20 08:39:08 crc kubenswrapper[4948]: I0220 08:39:08.300900 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="df133c52e6c7c766068e5890945bf88c2349bad99b1086eddf879822786cef6f" exitCode=0 Feb 20 08:39:08 crc kubenswrapper[4948]: I0220 08:39:08.301012 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"df133c52e6c7c766068e5890945bf88c2349bad99b1086eddf879822786cef6f"} Feb 20 08:39:08 crc kubenswrapper[4948]: I0220 08:39:08.301515 4948 scope.go:117] "RemoveContainer" containerID="abbd3a7ae1687f175b35d7d4eb71a990e4917e9020e7309ea75d7aa3bd6e4e43" Feb 20 08:39:09 crc kubenswrapper[4948]: I0220 08:39:09.317471 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707"} Feb 20 08:41:08 crc kubenswrapper[4948]: I0220 08:41:08.024850 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:41:08 crc kubenswrapper[4948]: I0220 08:41:08.025488 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:41:38 crc kubenswrapper[4948]: I0220 08:41:38.025140 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:41:38 crc kubenswrapper[4948]: I0220 08:41:38.025726 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:41:56 crc kubenswrapper[4948]: I0220 08:41:56.097322 4948 generic.go:334] "Generic (PLEG): container finished" podID="6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" containerID="e26f1bc45952ca3a3414ed06ca781d747c45268c9f970dd0c34eb0f719572756" exitCode=0 Feb 20 08:41:56 crc kubenswrapper[4948]: I0220 08:41:56.097478 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" event={"ID":"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a","Type":"ContainerDied","Data":"e26f1bc45952ca3a3414ed06ca781d747c45268c9f970dd0c34eb0f719572756"} Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.636697 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.790206 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-secret-0\") pod \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.790406 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-ssh-key-openstack-edpm-ipam\") pod \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.790543 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-combined-ca-bundle\") pod \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.790582 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8gnt\" (UniqueName: \"kubernetes.io/projected/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-kube-api-access-x8gnt\") pod \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.790624 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-inventory\") pod \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\" (UID: \"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a\") " Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.798178 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" (UID: "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.798252 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-kube-api-access-x8gnt" (OuterVolumeSpecName: "kube-api-access-x8gnt") pod "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" (UID: "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a"). InnerVolumeSpecName "kube-api-access-x8gnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.824493 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-inventory" (OuterVolumeSpecName: "inventory") pod "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" (UID: "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.830669 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" (UID: "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.831233 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" (UID: "6e619eac-dfdc-4c8c-aa51-fc08ba58b49a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.896162 4948 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.896240 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.896265 4948 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.896283 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8gnt\" (UniqueName: \"kubernetes.io/projected/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-kube-api-access-x8gnt\") on node \"crc\" DevicePath \"\"" Feb 20 08:41:57 crc kubenswrapper[4948]: I0220 08:41:57.896333 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e619eac-dfdc-4c8c-aa51-fc08ba58b49a-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.121847 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" event={"ID":"6e619eac-dfdc-4c8c-aa51-fc08ba58b49a","Type":"ContainerDied","Data":"bed584d1cffc5c44267aa8da79eb83c0547d67f904d60a696d0a5f0983625d4a"} Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.121948 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bed584d1cffc5c44267aa8da79eb83c0547d67f904d60a696d0a5f0983625d4a" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.122147 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.260311 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr"] Feb 20 08:41:58 crc kubenswrapper[4948]: E0220 08:41:58.261243 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.261262 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.261551 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e619eac-dfdc-4c8c-aa51-fc08ba58b49a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.262318 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.265195 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.266067 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.266297 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.267080 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.267293 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.267429 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.267832 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.277524 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr"] Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.404755 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.405089 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.405216 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.405361 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.405464 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.405607 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ljgt\" (UniqueName: \"kubernetes.io/projected/6005ded0-9e21-4908-bbf0-33c710ba4341-kube-api-access-8ljgt\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.405739 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.405902 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.406081 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.406181 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.406266 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.508465 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.508638 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.508681 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.508760 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.508815 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.508853 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ljgt\" (UniqueName: \"kubernetes.io/projected/6005ded0-9e21-4908-bbf0-33c710ba4341-kube-api-access-8ljgt\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.508926 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.509051 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.509097 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.509137 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.509177 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.510208 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.512362 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.512945 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.513774 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.514102 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.515300 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.515452 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.516161 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.530702 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.533374 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.537705 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ljgt\" (UniqueName: \"kubernetes.io/projected/6005ded0-9e21-4908-bbf0-33c710ba4341-kube-api-access-8ljgt\") pod \"nova-edpm-deployment-openstack-edpm-ipam-p22xr\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:58 crc kubenswrapper[4948]: I0220 08:41:58.589637 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:41:59 crc kubenswrapper[4948]: I0220 08:41:59.196170 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr"] Feb 20 08:42:00 crc kubenswrapper[4948]: I0220 08:42:00.154928 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" event={"ID":"6005ded0-9e21-4908-bbf0-33c710ba4341","Type":"ContainerStarted","Data":"2f2f83e7b2bae4f94ae51a61eb3c4783c5606e273956b4e39c8926e563dd497a"} Feb 20 08:42:00 crc kubenswrapper[4948]: I0220 08:42:00.155507 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" event={"ID":"6005ded0-9e21-4908-bbf0-33c710ba4341","Type":"ContainerStarted","Data":"f847f6d88e05df089793a5fb35fc239a9b46a7a0da87f3039554ac3950729267"} Feb 20 08:42:00 crc kubenswrapper[4948]: I0220 08:42:00.187950 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" podStartSLOduration=1.693300743 podStartE2EDuration="2.187931383s" podCreationTimestamp="2026-02-20 08:41:58 +0000 UTC" firstStartedPulling="2026-02-20 08:41:59.206956337 +0000 UTC m=+2168.181451157" lastFinishedPulling="2026-02-20 08:41:59.701586947 +0000 UTC m=+2168.676081797" observedRunningTime="2026-02-20 08:42:00.177130476 +0000 UTC m=+2169.151625316" watchObservedRunningTime="2026-02-20 08:42:00.187931383 +0000 UTC m=+2169.162426203" Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.024731 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.025352 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.025405 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.026300 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.026367 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" gracePeriod=600 Feb 20 08:42:08 crc kubenswrapper[4948]: E0220 08:42:08.157321 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.268392 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" exitCode=0 Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.268468 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707"} Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.268508 4948 scope.go:117] "RemoveContainer" containerID="df133c52e6c7c766068e5890945bf88c2349bad99b1086eddf879822786cef6f" Feb 20 08:42:08 crc kubenswrapper[4948]: I0220 08:42:08.271408 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:42:08 crc kubenswrapper[4948]: E0220 08:42:08.272195 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:42:22 crc kubenswrapper[4948]: I0220 08:42:22.725185 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:42:22 crc kubenswrapper[4948]: E0220 08:42:22.726115 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:42:37 crc kubenswrapper[4948]: I0220 08:42:37.729556 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:42:37 crc kubenswrapper[4948]: E0220 08:42:37.730225 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:42:42 crc kubenswrapper[4948]: I0220 08:42:42.802818 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-88wr7"] Feb 20 08:42:42 crc kubenswrapper[4948]: I0220 08:42:42.808551 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:42 crc kubenswrapper[4948]: I0220 08:42:42.822364 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-88wr7"] Feb 20 08:42:42 crc kubenswrapper[4948]: I0220 08:42:42.917512 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-catalog-content\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:42 crc kubenswrapper[4948]: I0220 08:42:42.917581 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg2dj\" (UniqueName: \"kubernetes.io/projected/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-kube-api-access-gg2dj\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:42 crc kubenswrapper[4948]: I0220 08:42:42.917633 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-utilities\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:43 crc kubenswrapper[4948]: I0220 08:42:43.019538 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg2dj\" (UniqueName: \"kubernetes.io/projected/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-kube-api-access-gg2dj\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:43 crc kubenswrapper[4948]: I0220 08:42:43.019599 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-utilities\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:43 crc kubenswrapper[4948]: I0220 08:42:43.019737 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-catalog-content\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:43 crc kubenswrapper[4948]: I0220 08:42:43.020268 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-catalog-content\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:43 crc kubenswrapper[4948]: I0220 08:42:43.020524 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-utilities\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:43 crc kubenswrapper[4948]: I0220 08:42:43.046744 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg2dj\" (UniqueName: \"kubernetes.io/projected/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-kube-api-access-gg2dj\") pod \"redhat-operators-88wr7\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:43 crc kubenswrapper[4948]: I0220 08:42:43.149288 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:43 crc kubenswrapper[4948]: I0220 08:42:43.640378 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-88wr7"] Feb 20 08:42:44 crc kubenswrapper[4948]: I0220 08:42:44.633670 4948 generic.go:334] "Generic (PLEG): container finished" podID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerID="020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca" exitCode=0 Feb 20 08:42:44 crc kubenswrapper[4948]: I0220 08:42:44.633809 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88wr7" event={"ID":"4bc1b6a7-dfaf-4c10-88be-f482a5c58667","Type":"ContainerDied","Data":"020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca"} Feb 20 08:42:44 crc kubenswrapper[4948]: I0220 08:42:44.634011 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88wr7" event={"ID":"4bc1b6a7-dfaf-4c10-88be-f482a5c58667","Type":"ContainerStarted","Data":"5c209d2da2d8783cbae32471f9af6f43d7c3ec2277c0cd928735be812a4502bc"} Feb 20 08:42:45 crc kubenswrapper[4948]: I0220 08:42:45.649810 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88wr7" event={"ID":"4bc1b6a7-dfaf-4c10-88be-f482a5c58667","Type":"ContainerStarted","Data":"7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73"} Feb 20 08:42:46 crc kubenswrapper[4948]: I0220 08:42:46.667738 4948 generic.go:334] "Generic (PLEG): container finished" podID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerID="7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73" exitCode=0 Feb 20 08:42:46 crc kubenswrapper[4948]: I0220 08:42:46.667832 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88wr7" event={"ID":"4bc1b6a7-dfaf-4c10-88be-f482a5c58667","Type":"ContainerDied","Data":"7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73"} Feb 20 08:42:47 crc kubenswrapper[4948]: I0220 08:42:47.680801 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88wr7" event={"ID":"4bc1b6a7-dfaf-4c10-88be-f482a5c58667","Type":"ContainerStarted","Data":"f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d"} Feb 20 08:42:51 crc kubenswrapper[4948]: I0220 08:42:51.735670 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:42:51 crc kubenswrapper[4948]: E0220 08:42:51.736961 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:42:53 crc kubenswrapper[4948]: I0220 08:42:53.149532 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:53 crc kubenswrapper[4948]: I0220 08:42:53.149576 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:42:54 crc kubenswrapper[4948]: I0220 08:42:54.220637 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-88wr7" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="registry-server" probeResult="failure" output=< Feb 20 08:42:54 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 08:42:54 crc kubenswrapper[4948]: > Feb 20 08:43:03 crc kubenswrapper[4948]: I0220 08:43:03.241590 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:43:03 crc kubenswrapper[4948]: I0220 08:43:03.268725 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-88wr7" podStartSLOduration=18.807062757 podStartE2EDuration="21.268694457s" podCreationTimestamp="2026-02-20 08:42:42 +0000 UTC" firstStartedPulling="2026-02-20 08:42:44.636705196 +0000 UTC m=+2213.611200056" lastFinishedPulling="2026-02-20 08:42:47.098336896 +0000 UTC m=+2216.072831756" observedRunningTime="2026-02-20 08:42:47.708618342 +0000 UTC m=+2216.683113172" watchObservedRunningTime="2026-02-20 08:43:03.268694457 +0000 UTC m=+2232.243189317" Feb 20 08:43:03 crc kubenswrapper[4948]: I0220 08:43:03.309125 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:43:03 crc kubenswrapper[4948]: I0220 08:43:03.484861 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-88wr7"] Feb 20 08:43:04 crc kubenswrapper[4948]: I0220 08:43:04.869772 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-88wr7" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="registry-server" containerID="cri-o://f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d" gracePeriod=2 Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.298297 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.450722 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-catalog-content\") pod \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.450829 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-utilities\") pod \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.451157 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gg2dj\" (UniqueName: \"kubernetes.io/projected/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-kube-api-access-gg2dj\") pod \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\" (UID: \"4bc1b6a7-dfaf-4c10-88be-f482a5c58667\") " Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.451919 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-utilities" (OuterVolumeSpecName: "utilities") pod "4bc1b6a7-dfaf-4c10-88be-f482a5c58667" (UID: "4bc1b6a7-dfaf-4c10-88be-f482a5c58667"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.457195 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-kube-api-access-gg2dj" (OuterVolumeSpecName: "kube-api-access-gg2dj") pod "4bc1b6a7-dfaf-4c10-88be-f482a5c58667" (UID: "4bc1b6a7-dfaf-4c10-88be-f482a5c58667"). InnerVolumeSpecName "kube-api-access-gg2dj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.553788 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.553833 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gg2dj\" (UniqueName: \"kubernetes.io/projected/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-kube-api-access-gg2dj\") on node \"crc\" DevicePath \"\"" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.567752 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4bc1b6a7-dfaf-4c10-88be-f482a5c58667" (UID: "4bc1b6a7-dfaf-4c10-88be-f482a5c58667"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.655291 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bc1b6a7-dfaf-4c10-88be-f482a5c58667-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.881219 4948 generic.go:334] "Generic (PLEG): container finished" podID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerID="f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d" exitCode=0 Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.881280 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88wr7" event={"ID":"4bc1b6a7-dfaf-4c10-88be-f482a5c58667","Type":"ContainerDied","Data":"f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d"} Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.881309 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88wr7" event={"ID":"4bc1b6a7-dfaf-4c10-88be-f482a5c58667","Type":"ContainerDied","Data":"5c209d2da2d8783cbae32471f9af6f43d7c3ec2277c0cd928735be812a4502bc"} Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.881348 4948 scope.go:117] "RemoveContainer" containerID="f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.881481 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-88wr7" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.911538 4948 scope.go:117] "RemoveContainer" containerID="7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.913069 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-88wr7"] Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.920568 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-88wr7"] Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.938540 4948 scope.go:117] "RemoveContainer" containerID="020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.984565 4948 scope.go:117] "RemoveContainer" containerID="f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d" Feb 20 08:43:05 crc kubenswrapper[4948]: E0220 08:43:05.985222 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d\": container with ID starting with f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d not found: ID does not exist" containerID="f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.985290 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d"} err="failed to get container status \"f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d\": rpc error: code = NotFound desc = could not find container \"f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d\": container with ID starting with f9d71de10f48d2736cee59456b183da30aed2ce7e43fe91b4241385713ea2e1d not found: ID does not exist" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.985322 4948 scope.go:117] "RemoveContainer" containerID="7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73" Feb 20 08:43:05 crc kubenswrapper[4948]: E0220 08:43:05.985771 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73\": container with ID starting with 7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73 not found: ID does not exist" containerID="7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.985804 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73"} err="failed to get container status \"7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73\": rpc error: code = NotFound desc = could not find container \"7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73\": container with ID starting with 7ca67f0db7f8dd9db858f07ad294b0aa2364e71f346b50828413c9852296cb73 not found: ID does not exist" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.985828 4948 scope.go:117] "RemoveContainer" containerID="020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca" Feb 20 08:43:05 crc kubenswrapper[4948]: E0220 08:43:05.986088 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca\": container with ID starting with 020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca not found: ID does not exist" containerID="020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca" Feb 20 08:43:05 crc kubenswrapper[4948]: I0220 08:43:05.986121 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca"} err="failed to get container status \"020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca\": rpc error: code = NotFound desc = could not find container \"020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca\": container with ID starting with 020c49ea97fd0d2abaf92d2a4ecacff3dea8fe68fd14d2c0bda270118de765ca not found: ID does not exist" Feb 20 08:43:06 crc kubenswrapper[4948]: I0220 08:43:06.722424 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:43:06 crc kubenswrapper[4948]: E0220 08:43:06.722758 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:43:07 crc kubenswrapper[4948]: I0220 08:43:07.737619 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" path="/var/lib/kubelet/pods/4bc1b6a7-dfaf-4c10-88be-f482a5c58667/volumes" Feb 20 08:43:17 crc kubenswrapper[4948]: I0220 08:43:17.723619 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:43:17 crc kubenswrapper[4948]: E0220 08:43:17.725308 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:43:29 crc kubenswrapper[4948]: I0220 08:43:29.722902 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:43:29 crc kubenswrapper[4948]: E0220 08:43:29.723587 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:43:44 crc kubenswrapper[4948]: I0220 08:43:44.722986 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:43:44 crc kubenswrapper[4948]: E0220 08:43:44.723609 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:43:56 crc kubenswrapper[4948]: I0220 08:43:56.723511 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:43:56 crc kubenswrapper[4948]: E0220 08:43:56.724840 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.015366 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vqtpd"] Feb 20 08:43:59 crc kubenswrapper[4948]: E0220 08:43:59.016362 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="registry-server" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.016384 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="registry-server" Feb 20 08:43:59 crc kubenswrapper[4948]: E0220 08:43:59.016432 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="extract-content" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.016444 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="extract-content" Feb 20 08:43:59 crc kubenswrapper[4948]: E0220 08:43:59.016465 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="extract-utilities" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.016477 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="extract-utilities" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.016804 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bc1b6a7-dfaf-4c10-88be-f482a5c58667" containerName="registry-server" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.018779 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.022432 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vqtpd"] Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.181062 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sctnn\" (UniqueName: \"kubernetes.io/projected/8aaa9295-db53-415c-a01d-80328833af1c-kube-api-access-sctnn\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.181154 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-catalog-content\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.181262 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-utilities\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.283298 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-utilities\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.283404 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sctnn\" (UniqueName: \"kubernetes.io/projected/8aaa9295-db53-415c-a01d-80328833af1c-kube-api-access-sctnn\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.283469 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-catalog-content\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.283994 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-catalog-content\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.284000 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-utilities\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.305828 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sctnn\" (UniqueName: \"kubernetes.io/projected/8aaa9295-db53-415c-a01d-80328833af1c-kube-api-access-sctnn\") pod \"community-operators-vqtpd\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.401062 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:43:59 crc kubenswrapper[4948]: I0220 08:43:59.928068 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vqtpd"] Feb 20 08:44:00 crc kubenswrapper[4948]: I0220 08:44:00.497833 4948 generic.go:334] "Generic (PLEG): container finished" podID="8aaa9295-db53-415c-a01d-80328833af1c" containerID="04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8" exitCode=0 Feb 20 08:44:00 crc kubenswrapper[4948]: I0220 08:44:00.497934 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqtpd" event={"ID":"8aaa9295-db53-415c-a01d-80328833af1c","Type":"ContainerDied","Data":"04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8"} Feb 20 08:44:00 crc kubenswrapper[4948]: I0220 08:44:00.498410 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqtpd" event={"ID":"8aaa9295-db53-415c-a01d-80328833af1c","Type":"ContainerStarted","Data":"35e21374dff7e8f3d55b2da70cd138e5ac76100c8db55005f609ee420971b833"} Feb 20 08:44:00 crc kubenswrapper[4948]: I0220 08:44:00.501557 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:44:02 crc kubenswrapper[4948]: I0220 08:44:02.524201 4948 generic.go:334] "Generic (PLEG): container finished" podID="8aaa9295-db53-415c-a01d-80328833af1c" containerID="6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2" exitCode=0 Feb 20 08:44:02 crc kubenswrapper[4948]: I0220 08:44:02.524271 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqtpd" event={"ID":"8aaa9295-db53-415c-a01d-80328833af1c","Type":"ContainerDied","Data":"6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2"} Feb 20 08:44:03 crc kubenswrapper[4948]: I0220 08:44:03.538082 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqtpd" event={"ID":"8aaa9295-db53-415c-a01d-80328833af1c","Type":"ContainerStarted","Data":"1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732"} Feb 20 08:44:03 crc kubenswrapper[4948]: I0220 08:44:03.577632 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vqtpd" podStartSLOduration=3.069184415 podStartE2EDuration="5.577610884s" podCreationTimestamp="2026-02-20 08:43:58 +0000 UTC" firstStartedPulling="2026-02-20 08:44:00.500778716 +0000 UTC m=+2289.475273586" lastFinishedPulling="2026-02-20 08:44:03.009205195 +0000 UTC m=+2291.983700055" observedRunningTime="2026-02-20 08:44:03.568305543 +0000 UTC m=+2292.542800403" watchObservedRunningTime="2026-02-20 08:44:03.577610884 +0000 UTC m=+2292.552105714" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.719123 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-52bhl"] Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.721554 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.754247 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-52bhl"] Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.894718 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chz94\" (UniqueName: \"kubernetes.io/projected/7e320036-f16c-4305-8591-c2f37e560ae1-kube-api-access-chz94\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.894784 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-utilities\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.894833 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-catalog-content\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.997194 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-utilities\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.997284 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-catalog-content\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.997632 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chz94\" (UniqueName: \"kubernetes.io/projected/7e320036-f16c-4305-8591-c2f37e560ae1-kube-api-access-chz94\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.997780 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-catalog-content\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:07 crc kubenswrapper[4948]: I0220 08:44:07.997875 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-utilities\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:08 crc kubenswrapper[4948]: I0220 08:44:08.027453 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chz94\" (UniqueName: \"kubernetes.io/projected/7e320036-f16c-4305-8591-c2f37e560ae1-kube-api-access-chz94\") pod \"certified-operators-52bhl\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:08 crc kubenswrapper[4948]: I0220 08:44:08.057277 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:08 crc kubenswrapper[4948]: I0220 08:44:08.588889 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-52bhl"] Feb 20 08:44:08 crc kubenswrapper[4948]: W0220 08:44:08.602702 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e320036_f16c_4305_8591_c2f37e560ae1.slice/crio-36c7b6d8b8c2c363ef13c4f00576bc6b814e0270caaf06486e95056f5e0456ca WatchSource:0}: Error finding container 36c7b6d8b8c2c363ef13c4f00576bc6b814e0270caaf06486e95056f5e0456ca: Status 404 returned error can't find the container with id 36c7b6d8b8c2c363ef13c4f00576bc6b814e0270caaf06486e95056f5e0456ca Feb 20 08:44:09 crc kubenswrapper[4948]: I0220 08:44:09.401421 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:44:09 crc kubenswrapper[4948]: I0220 08:44:09.401825 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:44:09 crc kubenswrapper[4948]: I0220 08:44:09.468183 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:44:09 crc kubenswrapper[4948]: I0220 08:44:09.602952 4948 generic.go:334] "Generic (PLEG): container finished" podID="7e320036-f16c-4305-8591-c2f37e560ae1" containerID="b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2" exitCode=0 Feb 20 08:44:09 crc kubenswrapper[4948]: I0220 08:44:09.606222 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52bhl" event={"ID":"7e320036-f16c-4305-8591-c2f37e560ae1","Type":"ContainerDied","Data":"b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2"} Feb 20 08:44:09 crc kubenswrapper[4948]: I0220 08:44:09.606282 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52bhl" event={"ID":"7e320036-f16c-4305-8591-c2f37e560ae1","Type":"ContainerStarted","Data":"36c7b6d8b8c2c363ef13c4f00576bc6b814e0270caaf06486e95056f5e0456ca"} Feb 20 08:44:09 crc kubenswrapper[4948]: I0220 08:44:09.671254 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:44:09 crc kubenswrapper[4948]: I0220 08:44:09.723145 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:44:09 crc kubenswrapper[4948]: E0220 08:44:09.723483 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:44:10 crc kubenswrapper[4948]: I0220 08:44:10.612352 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52bhl" event={"ID":"7e320036-f16c-4305-8591-c2f37e560ae1","Type":"ContainerStarted","Data":"183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6"} Feb 20 08:44:11 crc kubenswrapper[4948]: I0220 08:44:11.645382 4948 generic.go:334] "Generic (PLEG): container finished" podID="7e320036-f16c-4305-8591-c2f37e560ae1" containerID="183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6" exitCode=0 Feb 20 08:44:11 crc kubenswrapper[4948]: I0220 08:44:11.645440 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52bhl" event={"ID":"7e320036-f16c-4305-8591-c2f37e560ae1","Type":"ContainerDied","Data":"183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6"} Feb 20 08:44:11 crc kubenswrapper[4948]: I0220 08:44:11.893545 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vqtpd"] Feb 20 08:44:11 crc kubenswrapper[4948]: I0220 08:44:11.894105 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vqtpd" podUID="8aaa9295-db53-415c-a01d-80328833af1c" containerName="registry-server" containerID="cri-o://1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732" gracePeriod=2 Feb 20 08:44:12 crc kubenswrapper[4948]: I0220 08:44:12.655454 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52bhl" event={"ID":"7e320036-f16c-4305-8591-c2f37e560ae1","Type":"ContainerStarted","Data":"8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036"} Feb 20 08:44:12 crc kubenswrapper[4948]: I0220 08:44:12.679539 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-52bhl" podStartSLOduration=3.245298351 podStartE2EDuration="5.679512661s" podCreationTimestamp="2026-02-20 08:44:07 +0000 UTC" firstStartedPulling="2026-02-20 08:44:09.609229686 +0000 UTC m=+2298.583724526" lastFinishedPulling="2026-02-20 08:44:12.043444016 +0000 UTC m=+2301.017938836" observedRunningTime="2026-02-20 08:44:12.673154614 +0000 UTC m=+2301.647649444" watchObservedRunningTime="2026-02-20 08:44:12.679512661 +0000 UTC m=+2301.654007481" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.569205 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.666147 4948 generic.go:334] "Generic (PLEG): container finished" podID="8aaa9295-db53-415c-a01d-80328833af1c" containerID="1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732" exitCode=0 Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.666573 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqtpd" event={"ID":"8aaa9295-db53-415c-a01d-80328833af1c","Type":"ContainerDied","Data":"1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732"} Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.666676 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqtpd" event={"ID":"8aaa9295-db53-415c-a01d-80328833af1c","Type":"ContainerDied","Data":"35e21374dff7e8f3d55b2da70cd138e5ac76100c8db55005f609ee420971b833"} Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.666695 4948 scope.go:117] "RemoveContainer" containerID="1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.666779 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqtpd" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.705656 4948 scope.go:117] "RemoveContainer" containerID="6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.715306 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sctnn\" (UniqueName: \"kubernetes.io/projected/8aaa9295-db53-415c-a01d-80328833af1c-kube-api-access-sctnn\") pod \"8aaa9295-db53-415c-a01d-80328833af1c\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.715498 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-utilities\") pod \"8aaa9295-db53-415c-a01d-80328833af1c\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.715636 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-catalog-content\") pod \"8aaa9295-db53-415c-a01d-80328833af1c\" (UID: \"8aaa9295-db53-415c-a01d-80328833af1c\") " Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.721834 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-utilities" (OuterVolumeSpecName: "utilities") pod "8aaa9295-db53-415c-a01d-80328833af1c" (UID: "8aaa9295-db53-415c-a01d-80328833af1c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.725227 4948 scope.go:117] "RemoveContainer" containerID="04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.727216 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aaa9295-db53-415c-a01d-80328833af1c-kube-api-access-sctnn" (OuterVolumeSpecName: "kube-api-access-sctnn") pod "8aaa9295-db53-415c-a01d-80328833af1c" (UID: "8aaa9295-db53-415c-a01d-80328833af1c"). InnerVolumeSpecName "kube-api-access-sctnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.775077 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8aaa9295-db53-415c-a01d-80328833af1c" (UID: "8aaa9295-db53-415c-a01d-80328833af1c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.818405 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sctnn\" (UniqueName: \"kubernetes.io/projected/8aaa9295-db53-415c-a01d-80328833af1c-kube-api-access-sctnn\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.818440 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.818449 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aaa9295-db53-415c-a01d-80328833af1c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.821090 4948 scope.go:117] "RemoveContainer" containerID="1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732" Feb 20 08:44:13 crc kubenswrapper[4948]: E0220 08:44:13.821760 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732\": container with ID starting with 1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732 not found: ID does not exist" containerID="1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.821801 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732"} err="failed to get container status \"1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732\": rpc error: code = NotFound desc = could not find container \"1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732\": container with ID starting with 1c1738fd62956b1cf907c7bb63dad6c36c4b8e5f2086bdcaa6aedd2a9f973732 not found: ID does not exist" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.821826 4948 scope.go:117] "RemoveContainer" containerID="6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2" Feb 20 08:44:13 crc kubenswrapper[4948]: E0220 08:44:13.822107 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2\": container with ID starting with 6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2 not found: ID does not exist" containerID="6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.822145 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2"} err="failed to get container status \"6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2\": rpc error: code = NotFound desc = could not find container \"6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2\": container with ID starting with 6043e2e2a61ed40b0ced1f92a7c96922825a7177951ad0332aa05a8120ec9df2 not found: ID does not exist" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.822169 4948 scope.go:117] "RemoveContainer" containerID="04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8" Feb 20 08:44:13 crc kubenswrapper[4948]: E0220 08:44:13.822412 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8\": container with ID starting with 04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8 not found: ID does not exist" containerID="04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.822434 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8"} err="failed to get container status \"04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8\": rpc error: code = NotFound desc = could not find container \"04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8\": container with ID starting with 04150064869f7e2104b08b6a1ea4042774631d5a1e64200e7d0b648e85fc12c8 not found: ID does not exist" Feb 20 08:44:13 crc kubenswrapper[4948]: I0220 08:44:13.996270 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vqtpd"] Feb 20 08:44:14 crc kubenswrapper[4948]: I0220 08:44:14.003911 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vqtpd"] Feb 20 08:44:15 crc kubenswrapper[4948]: I0220 08:44:15.741586 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8aaa9295-db53-415c-a01d-80328833af1c" path="/var/lib/kubelet/pods/8aaa9295-db53-415c-a01d-80328833af1c/volumes" Feb 20 08:44:18 crc kubenswrapper[4948]: I0220 08:44:18.058468 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:18 crc kubenswrapper[4948]: I0220 08:44:18.058756 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:18 crc kubenswrapper[4948]: I0220 08:44:18.133099 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:18 crc kubenswrapper[4948]: I0220 08:44:18.779966 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:19 crc kubenswrapper[4948]: I0220 08:44:19.296639 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-52bhl"] Feb 20 08:44:20 crc kubenswrapper[4948]: I0220 08:44:20.750641 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-52bhl" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" containerName="registry-server" containerID="cri-o://8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036" gracePeriod=2 Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.197570 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.382533 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-utilities\") pod \"7e320036-f16c-4305-8591-c2f37e560ae1\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.382588 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-catalog-content\") pod \"7e320036-f16c-4305-8591-c2f37e560ae1\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.382795 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chz94\" (UniqueName: \"kubernetes.io/projected/7e320036-f16c-4305-8591-c2f37e560ae1-kube-api-access-chz94\") pod \"7e320036-f16c-4305-8591-c2f37e560ae1\" (UID: \"7e320036-f16c-4305-8591-c2f37e560ae1\") " Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.383806 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-utilities" (OuterVolumeSpecName: "utilities") pod "7e320036-f16c-4305-8591-c2f37e560ae1" (UID: "7e320036-f16c-4305-8591-c2f37e560ae1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.389246 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e320036-f16c-4305-8591-c2f37e560ae1-kube-api-access-chz94" (OuterVolumeSpecName: "kube-api-access-chz94") pod "7e320036-f16c-4305-8591-c2f37e560ae1" (UID: "7e320036-f16c-4305-8591-c2f37e560ae1"). InnerVolumeSpecName "kube-api-access-chz94". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.444021 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e320036-f16c-4305-8591-c2f37e560ae1" (UID: "7e320036-f16c-4305-8591-c2f37e560ae1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.485278 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chz94\" (UniqueName: \"kubernetes.io/projected/7e320036-f16c-4305-8591-c2f37e560ae1-kube-api-access-chz94\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.485321 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.485336 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e320036-f16c-4305-8591-c2f37e560ae1-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.732398 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:44:21 crc kubenswrapper[4948]: E0220 08:44:21.732915 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.764596 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-52bhl" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.764640 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52bhl" event={"ID":"7e320036-f16c-4305-8591-c2f37e560ae1","Type":"ContainerDied","Data":"8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036"} Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.764692 4948 scope.go:117] "RemoveContainer" containerID="8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.764539 4948 generic.go:334] "Generic (PLEG): container finished" podID="7e320036-f16c-4305-8591-c2f37e560ae1" containerID="8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036" exitCode=0 Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.767544 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52bhl" event={"ID":"7e320036-f16c-4305-8591-c2f37e560ae1","Type":"ContainerDied","Data":"36c7b6d8b8c2c363ef13c4f00576bc6b814e0270caaf06486e95056f5e0456ca"} Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.798501 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-52bhl"] Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.798544 4948 scope.go:117] "RemoveContainer" containerID="183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.807296 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-52bhl"] Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.829646 4948 scope.go:117] "RemoveContainer" containerID="b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.873201 4948 scope.go:117] "RemoveContainer" containerID="8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036" Feb 20 08:44:21 crc kubenswrapper[4948]: E0220 08:44:21.873765 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036\": container with ID starting with 8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036 not found: ID does not exist" containerID="8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.873817 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036"} err="failed to get container status \"8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036\": rpc error: code = NotFound desc = could not find container \"8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036\": container with ID starting with 8dfd7a07c4d2e10fac00ea5e02e285c69b8c28107d28b2e8147168697f405036 not found: ID does not exist" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.873851 4948 scope.go:117] "RemoveContainer" containerID="183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6" Feb 20 08:44:21 crc kubenswrapper[4948]: E0220 08:44:21.874244 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6\": container with ID starting with 183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6 not found: ID does not exist" containerID="183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.874269 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6"} err="failed to get container status \"183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6\": rpc error: code = NotFound desc = could not find container \"183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6\": container with ID starting with 183389c9a096dfb452e747af5b860f85837fe8aad0eeae056715142ef6fa86d6 not found: ID does not exist" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.874283 4948 scope.go:117] "RemoveContainer" containerID="b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2" Feb 20 08:44:21 crc kubenswrapper[4948]: E0220 08:44:21.874902 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2\": container with ID starting with b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2 not found: ID does not exist" containerID="b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2" Feb 20 08:44:21 crc kubenswrapper[4948]: I0220 08:44:21.874942 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2"} err="failed to get container status \"b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2\": rpc error: code = NotFound desc = could not find container \"b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2\": container with ID starting with b6344b3659b55c85ab17bb84f2d75f533dcdaf801d948da41f4dce7365aba5b2 not found: ID does not exist" Feb 20 08:44:23 crc kubenswrapper[4948]: I0220 08:44:23.733961 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" path="/var/lib/kubelet/pods/7e320036-f16c-4305-8591-c2f37e560ae1/volumes" Feb 20 08:44:27 crc kubenswrapper[4948]: I0220 08:44:27.819833 4948 generic.go:334] "Generic (PLEG): container finished" podID="6005ded0-9e21-4908-bbf0-33c710ba4341" containerID="2f2f83e7b2bae4f94ae51a61eb3c4783c5606e273956b4e39c8926e563dd497a" exitCode=0 Feb 20 08:44:27 crc kubenswrapper[4948]: I0220 08:44:27.819914 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" event={"ID":"6005ded0-9e21-4908-bbf0-33c710ba4341","Type":"ContainerDied","Data":"2f2f83e7b2bae4f94ae51a61eb3c4783c5606e273956b4e39c8926e563dd497a"} Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.207805 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.352712 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-extra-config-0\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.352751 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-1\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.352778 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-0\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.352848 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ljgt\" (UniqueName: \"kubernetes.io/projected/6005ded0-9e21-4908-bbf0-33c710ba4341-kube-api-access-8ljgt\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.352880 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-2\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.352917 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-3\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.352972 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-inventory\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.353027 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-ssh-key-openstack-edpm-ipam\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.353075 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-combined-ca-bundle\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.353132 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-1\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.353173 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-0\") pod \"6005ded0-9e21-4908-bbf0-33c710ba4341\" (UID: \"6005ded0-9e21-4908-bbf0-33c710ba4341\") " Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.358377 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.358805 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6005ded0-9e21-4908-bbf0-33c710ba4341-kube-api-access-8ljgt" (OuterVolumeSpecName: "kube-api-access-8ljgt") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "kube-api-access-8ljgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.386217 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.388284 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.390475 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.391788 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.398135 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.398884 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.400072 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.400287 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.422009 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-inventory" (OuterVolumeSpecName: "inventory") pod "6005ded0-9e21-4908-bbf0-33c710ba4341" (UID: "6005ded0-9e21-4908-bbf0-33c710ba4341"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.455711 4948 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456090 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ljgt\" (UniqueName: \"kubernetes.io/projected/6005ded0-9e21-4908-bbf0-33c710ba4341-kube-api-access-8ljgt\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456104 4948 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456116 4948 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456129 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456141 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456152 4948 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456165 4948 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456176 4948 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456191 4948 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.456203 4948 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6005ded0-9e21-4908-bbf0-33c710ba4341-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.837003 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" event={"ID":"6005ded0-9e21-4908-bbf0-33c710ba4341","Type":"ContainerDied","Data":"f847f6d88e05df089793a5fb35fc239a9b46a7a0da87f3039554ac3950729267"} Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.837044 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f847f6d88e05df089793a5fb35fc239a9b46a7a0da87f3039554ac3950729267" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.837051 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-p22xr" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.950217 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd"] Feb 20 08:44:29 crc kubenswrapper[4948]: E0220 08:44:29.950641 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aaa9295-db53-415c-a01d-80328833af1c" containerName="registry-server" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.950665 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aaa9295-db53-415c-a01d-80328833af1c" containerName="registry-server" Feb 20 08:44:29 crc kubenswrapper[4948]: E0220 08:44:29.950679 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" containerName="extract-utilities" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.950686 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" containerName="extract-utilities" Feb 20 08:44:29 crc kubenswrapper[4948]: E0220 08:44:29.950698 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6005ded0-9e21-4908-bbf0-33c710ba4341" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.950706 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="6005ded0-9e21-4908-bbf0-33c710ba4341" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 20 08:44:29 crc kubenswrapper[4948]: E0220 08:44:29.950733 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" containerName="extract-content" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.950740 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" containerName="extract-content" Feb 20 08:44:29 crc kubenswrapper[4948]: E0220 08:44:29.950753 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aaa9295-db53-415c-a01d-80328833af1c" containerName="extract-utilities" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.950760 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aaa9295-db53-415c-a01d-80328833af1c" containerName="extract-utilities" Feb 20 08:44:29 crc kubenswrapper[4948]: E0220 08:44:29.950773 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aaa9295-db53-415c-a01d-80328833af1c" containerName="extract-content" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.950781 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aaa9295-db53-415c-a01d-80328833af1c" containerName="extract-content" Feb 20 08:44:29 crc kubenswrapper[4948]: E0220 08:44:29.950795 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" containerName="registry-server" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.950802 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" containerName="registry-server" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.951072 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e320036-f16c-4305-8591-c2f37e560ae1" containerName="registry-server" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.951090 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="6005ded0-9e21-4908-bbf0-33c710ba4341" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.951120 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aaa9295-db53-415c-a01d-80328833af1c" containerName="registry-server" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.951879 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.954019 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.954232 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.954356 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.954470 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sgx49" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.955171 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 08:44:29 crc kubenswrapper[4948]: I0220 08:44:29.959945 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd"] Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.065509 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp6fd\" (UniqueName: \"kubernetes.io/projected/06ed5839-bc31-4691-8586-cd42c0413006-kube-api-access-xp6fd\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.065578 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.065603 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.065633 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.065653 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.065765 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.065846 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.167430 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.167481 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.167538 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.167560 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.167632 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp6fd\" (UniqueName: \"kubernetes.io/projected/06ed5839-bc31-4691-8586-cd42c0413006-kube-api-access-xp6fd\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.167675 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.167694 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.175249 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.177595 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.178375 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.179031 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.183639 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.197623 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.205738 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp6fd\" (UniqueName: \"kubernetes.io/projected/06ed5839-bc31-4691-8586-cd42c0413006-kube-api-access-xp6fd\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.267514 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:44:30 crc kubenswrapper[4948]: I0220 08:44:30.853187 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd"] Feb 20 08:44:30 crc kubenswrapper[4948]: W0220 08:44:30.858873 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06ed5839_bc31_4691_8586_cd42c0413006.slice/crio-d456a9922236504054c207d73b58e4c58e11b56c6a928f1ce1badf2b7fbf1f58 WatchSource:0}: Error finding container d456a9922236504054c207d73b58e4c58e11b56c6a928f1ce1badf2b7fbf1f58: Status 404 returned error can't find the container with id d456a9922236504054c207d73b58e4c58e11b56c6a928f1ce1badf2b7fbf1f58 Feb 20 08:44:31 crc kubenswrapper[4948]: I0220 08:44:31.855663 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" event={"ID":"06ed5839-bc31-4691-8586-cd42c0413006","Type":"ContainerStarted","Data":"9f9449ceeed4dadc49534402fd6aefa2d193a1d53da8937c431d3883a09b7e37"} Feb 20 08:44:31 crc kubenswrapper[4948]: I0220 08:44:31.856034 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" event={"ID":"06ed5839-bc31-4691-8586-cd42c0413006","Type":"ContainerStarted","Data":"d456a9922236504054c207d73b58e4c58e11b56c6a928f1ce1badf2b7fbf1f58"} Feb 20 08:44:31 crc kubenswrapper[4948]: I0220 08:44:31.889971 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" podStartSLOduration=2.469003812 podStartE2EDuration="2.889945458s" podCreationTimestamp="2026-02-20 08:44:29 +0000 UTC" firstStartedPulling="2026-02-20 08:44:30.861749882 +0000 UTC m=+2319.836244722" lastFinishedPulling="2026-02-20 08:44:31.282691548 +0000 UTC m=+2320.257186368" observedRunningTime="2026-02-20 08:44:31.873701736 +0000 UTC m=+2320.848196556" watchObservedRunningTime="2026-02-20 08:44:31.889945458 +0000 UTC m=+2320.864440288" Feb 20 08:44:32 crc kubenswrapper[4948]: I0220 08:44:32.722470 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:44:32 crc kubenswrapper[4948]: E0220 08:44:32.722833 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:44:44 crc kubenswrapper[4948]: I0220 08:44:44.723270 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:44:44 crc kubenswrapper[4948]: E0220 08:44:44.724561 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:44:59 crc kubenswrapper[4948]: I0220 08:44:59.723576 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:44:59 crc kubenswrapper[4948]: E0220 08:44:59.724821 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.143717 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn"] Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.145079 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.148137 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.157927 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn"] Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.158891 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.279300 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b7070ce-cb29-445c-8f62-dd8b8db24431-config-volume\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.279384 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb977\" (UniqueName: \"kubernetes.io/projected/1b7070ce-cb29-445c-8f62-dd8b8db24431-kube-api-access-sb977\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.279491 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b7070ce-cb29-445c-8f62-dd8b8db24431-secret-volume\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.381808 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb977\" (UniqueName: \"kubernetes.io/projected/1b7070ce-cb29-445c-8f62-dd8b8db24431-kube-api-access-sb977\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.382431 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b7070ce-cb29-445c-8f62-dd8b8db24431-secret-volume\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.382685 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b7070ce-cb29-445c-8f62-dd8b8db24431-config-volume\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.384257 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b7070ce-cb29-445c-8f62-dd8b8db24431-config-volume\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.393447 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b7070ce-cb29-445c-8f62-dd8b8db24431-secret-volume\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.405592 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb977\" (UniqueName: \"kubernetes.io/projected/1b7070ce-cb29-445c-8f62-dd8b8db24431-kube-api-access-sb977\") pod \"collect-profiles-29526285-5lvgn\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.463680 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:00 crc kubenswrapper[4948]: I0220 08:45:00.958186 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn"] Feb 20 08:45:01 crc kubenswrapper[4948]: I0220 08:45:01.452611 4948 generic.go:334] "Generic (PLEG): container finished" podID="1b7070ce-cb29-445c-8f62-dd8b8db24431" containerID="0487ec6712045782955f0041f4583049c1ce40e10480a8b6d989918d4fa3552c" exitCode=0 Feb 20 08:45:01 crc kubenswrapper[4948]: I0220 08:45:01.452669 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" event={"ID":"1b7070ce-cb29-445c-8f62-dd8b8db24431","Type":"ContainerDied","Data":"0487ec6712045782955f0041f4583049c1ce40e10480a8b6d989918d4fa3552c"} Feb 20 08:45:01 crc kubenswrapper[4948]: I0220 08:45:01.452992 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" event={"ID":"1b7070ce-cb29-445c-8f62-dd8b8db24431","Type":"ContainerStarted","Data":"0f5ebdcd3104c719474346869a61ea8bbb58759dd8a0ed7585670e633c56ed16"} Feb 20 08:45:02 crc kubenswrapper[4948]: I0220 08:45:02.806025 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.006985 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb977\" (UniqueName: \"kubernetes.io/projected/1b7070ce-cb29-445c-8f62-dd8b8db24431-kube-api-access-sb977\") pod \"1b7070ce-cb29-445c-8f62-dd8b8db24431\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.007471 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b7070ce-cb29-445c-8f62-dd8b8db24431-config-volume\") pod \"1b7070ce-cb29-445c-8f62-dd8b8db24431\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.007694 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b7070ce-cb29-445c-8f62-dd8b8db24431-secret-volume\") pod \"1b7070ce-cb29-445c-8f62-dd8b8db24431\" (UID: \"1b7070ce-cb29-445c-8f62-dd8b8db24431\") " Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.007852 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b7070ce-cb29-445c-8f62-dd8b8db24431-config-volume" (OuterVolumeSpecName: "config-volume") pod "1b7070ce-cb29-445c-8f62-dd8b8db24431" (UID: "1b7070ce-cb29-445c-8f62-dd8b8db24431"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.008335 4948 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b7070ce-cb29-445c-8f62-dd8b8db24431-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.012430 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b7070ce-cb29-445c-8f62-dd8b8db24431-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1b7070ce-cb29-445c-8f62-dd8b8db24431" (UID: "1b7070ce-cb29-445c-8f62-dd8b8db24431"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.012677 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b7070ce-cb29-445c-8f62-dd8b8db24431-kube-api-access-sb977" (OuterVolumeSpecName: "kube-api-access-sb977") pod "1b7070ce-cb29-445c-8f62-dd8b8db24431" (UID: "1b7070ce-cb29-445c-8f62-dd8b8db24431"). InnerVolumeSpecName "kube-api-access-sb977". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.109551 4948 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b7070ce-cb29-445c-8f62-dd8b8db24431-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.109586 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb977\" (UniqueName: \"kubernetes.io/projected/1b7070ce-cb29-445c-8f62-dd8b8db24431-kube-api-access-sb977\") on node \"crc\" DevicePath \"\"" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.470889 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" event={"ID":"1b7070ce-cb29-445c-8f62-dd8b8db24431","Type":"ContainerDied","Data":"0f5ebdcd3104c719474346869a61ea8bbb58759dd8a0ed7585670e633c56ed16"} Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.471502 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f5ebdcd3104c719474346869a61ea8bbb58759dd8a0ed7585670e633c56ed16" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.470949 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-5lvgn" Feb 20 08:45:03 crc kubenswrapper[4948]: E0220 08:45:03.675648 4948 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b7070ce_cb29_445c_8f62_dd8b8db24431.slice/crio-0f5ebdcd3104c719474346869a61ea8bbb58759dd8a0ed7585670e633c56ed16\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b7070ce_cb29_445c_8f62_dd8b8db24431.slice\": RecentStats: unable to find data in memory cache]" Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.876027 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk"] Feb 20 08:45:03 crc kubenswrapper[4948]: I0220 08:45:03.884761 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526240-lwsqk"] Feb 20 08:45:05 crc kubenswrapper[4948]: I0220 08:45:05.739187 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f77016a6-cd35-49df-b6b5-65b4858b41c9" path="/var/lib/kubelet/pods/f77016a6-cd35-49df-b6b5-65b4858b41c9/volumes" Feb 20 08:45:14 crc kubenswrapper[4948]: I0220 08:45:14.722273 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:45:14 crc kubenswrapper[4948]: E0220 08:45:14.723212 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:45:25 crc kubenswrapper[4948]: I0220 08:45:25.722965 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:45:25 crc kubenswrapper[4948]: E0220 08:45:25.723881 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:45:39 crc kubenswrapper[4948]: I0220 08:45:39.724226 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:45:39 crc kubenswrapper[4948]: E0220 08:45:39.725631 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:45:52 crc kubenswrapper[4948]: I0220 08:45:52.723539 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:45:52 crc kubenswrapper[4948]: E0220 08:45:52.724341 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:46:02 crc kubenswrapper[4948]: I0220 08:46:02.276139 4948 scope.go:117] "RemoveContainer" containerID="0520ff9e3a8f12a5228a352b59373978c12d5f10bb51923b12eab15acf564086" Feb 20 08:46:03 crc kubenswrapper[4948]: I0220 08:46:03.722625 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:46:03 crc kubenswrapper[4948]: E0220 08:46:03.723300 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:46:16 crc kubenswrapper[4948]: I0220 08:46:16.723164 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:46:16 crc kubenswrapper[4948]: E0220 08:46:16.724421 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:46:28 crc kubenswrapper[4948]: I0220 08:46:28.723158 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:46:28 crc kubenswrapper[4948]: E0220 08:46:28.724181 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:46:40 crc kubenswrapper[4948]: I0220 08:46:40.723243 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:46:40 crc kubenswrapper[4948]: E0220 08:46:40.725182 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:46:53 crc kubenswrapper[4948]: I0220 08:46:53.722930 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:46:53 crc kubenswrapper[4948]: E0220 08:46:53.723897 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:47:03 crc kubenswrapper[4948]: I0220 08:47:03.718470 4948 generic.go:334] "Generic (PLEG): container finished" podID="06ed5839-bc31-4691-8586-cd42c0413006" containerID="9f9449ceeed4dadc49534402fd6aefa2d193a1d53da8937c431d3883a09b7e37" exitCode=0 Feb 20 08:47:03 crc kubenswrapper[4948]: I0220 08:47:03.718569 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" event={"ID":"06ed5839-bc31-4691-8586-cd42c0413006","Type":"ContainerDied","Data":"9f9449ceeed4dadc49534402fd6aefa2d193a1d53da8937c431d3883a09b7e37"} Feb 20 08:47:04 crc kubenswrapper[4948]: I0220 08:47:04.723773 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:47:04 crc kubenswrapper[4948]: E0220 08:47:04.725157 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.230304 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.267211 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ssh-key-openstack-edpm-ipam\") pod \"06ed5839-bc31-4691-8586-cd42c0413006\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.267331 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-2\") pod \"06ed5839-bc31-4691-8586-cd42c0413006\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.267420 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-0\") pod \"06ed5839-bc31-4691-8586-cd42c0413006\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.267617 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-1\") pod \"06ed5839-bc31-4691-8586-cd42c0413006\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.267693 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-telemetry-combined-ca-bundle\") pod \"06ed5839-bc31-4691-8586-cd42c0413006\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.267749 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-inventory\") pod \"06ed5839-bc31-4691-8586-cd42c0413006\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.267896 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xp6fd\" (UniqueName: \"kubernetes.io/projected/06ed5839-bc31-4691-8586-cd42c0413006-kube-api-access-xp6fd\") pod \"06ed5839-bc31-4691-8586-cd42c0413006\" (UID: \"06ed5839-bc31-4691-8586-cd42c0413006\") " Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.274196 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06ed5839-bc31-4691-8586-cd42c0413006-kube-api-access-xp6fd" (OuterVolumeSpecName: "kube-api-access-xp6fd") pod "06ed5839-bc31-4691-8586-cd42c0413006" (UID: "06ed5839-bc31-4691-8586-cd42c0413006"). InnerVolumeSpecName "kube-api-access-xp6fd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.275575 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "06ed5839-bc31-4691-8586-cd42c0413006" (UID: "06ed5839-bc31-4691-8586-cd42c0413006"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.299851 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "06ed5839-bc31-4691-8586-cd42c0413006" (UID: "06ed5839-bc31-4691-8586-cd42c0413006"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.305411 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "06ed5839-bc31-4691-8586-cd42c0413006" (UID: "06ed5839-bc31-4691-8586-cd42c0413006"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.315899 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "06ed5839-bc31-4691-8586-cd42c0413006" (UID: "06ed5839-bc31-4691-8586-cd42c0413006"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.324335 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "06ed5839-bc31-4691-8586-cd42c0413006" (UID: "06ed5839-bc31-4691-8586-cd42c0413006"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.329622 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-inventory" (OuterVolumeSpecName: "inventory") pod "06ed5839-bc31-4691-8586-cd42c0413006" (UID: "06ed5839-bc31-4691-8586-cd42c0413006"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.370544 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.370573 4948 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.370585 4948 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.370595 4948 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.370603 4948 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.370615 4948 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06ed5839-bc31-4691-8586-cd42c0413006-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.370623 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xp6fd\" (UniqueName: \"kubernetes.io/projected/06ed5839-bc31-4691-8586-cd42c0413006-kube-api-access-xp6fd\") on node \"crc\" DevicePath \"\"" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.746104 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" event={"ID":"06ed5839-bc31-4691-8586-cd42c0413006","Type":"ContainerDied","Data":"d456a9922236504054c207d73b58e4c58e11b56c6a928f1ce1badf2b7fbf1f58"} Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.746158 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d456a9922236504054c207d73b58e4c58e11b56c6a928f1ce1badf2b7fbf1f58" Feb 20 08:47:05 crc kubenswrapper[4948]: I0220 08:47:05.747217 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd" Feb 20 08:47:16 crc kubenswrapper[4948]: I0220 08:47:16.739436 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:47:17 crc kubenswrapper[4948]: I0220 08:47:17.910560 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"dc74db389f6b091911817b21cddd65e37098545ff6852fc34fb4458a2808d454"} Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.943047 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhz5"] Feb 20 08:47:46 crc kubenswrapper[4948]: E0220 08:47:46.945877 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06ed5839-bc31-4691-8586-cd42c0413006" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.946044 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="06ed5839-bc31-4691-8586-cd42c0413006" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 20 08:47:46 crc kubenswrapper[4948]: E0220 08:47:46.946153 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b7070ce-cb29-445c-8f62-dd8b8db24431" containerName="collect-profiles" Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.946267 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b7070ce-cb29-445c-8f62-dd8b8db24431" containerName="collect-profiles" Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.946752 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b7070ce-cb29-445c-8f62-dd8b8db24431" containerName="collect-profiles" Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.955764 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="06ed5839-bc31-4691-8586-cd42c0413006" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.957836 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.958331 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhz5"] Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.991708 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-utilities\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.994326 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-catalog-content\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:46 crc kubenswrapper[4948]: I0220 08:47:46.994415 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7p69\" (UniqueName: \"kubernetes.io/projected/10cca249-45b2-462c-a348-5881307f2f98-kube-api-access-l7p69\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:47 crc kubenswrapper[4948]: I0220 08:47:47.096370 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-utilities\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:47 crc kubenswrapper[4948]: I0220 08:47:47.096482 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-catalog-content\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:47 crc kubenswrapper[4948]: I0220 08:47:47.096510 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7p69\" (UniqueName: \"kubernetes.io/projected/10cca249-45b2-462c-a348-5881307f2f98-kube-api-access-l7p69\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:47 crc kubenswrapper[4948]: I0220 08:47:47.097151 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-catalog-content\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:47 crc kubenswrapper[4948]: I0220 08:47:47.097192 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-utilities\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:47 crc kubenswrapper[4948]: I0220 08:47:47.118864 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7p69\" (UniqueName: \"kubernetes.io/projected/10cca249-45b2-462c-a348-5881307f2f98-kube-api-access-l7p69\") pod \"redhat-marketplace-jbhz5\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:47 crc kubenswrapper[4948]: I0220 08:47:47.300411 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:47 crc kubenswrapper[4948]: I0220 08:47:47.858333 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhz5"] Feb 20 08:47:48 crc kubenswrapper[4948]: I0220 08:47:48.236133 4948 generic.go:334] "Generic (PLEG): container finished" podID="10cca249-45b2-462c-a348-5881307f2f98" containerID="ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137" exitCode=0 Feb 20 08:47:48 crc kubenswrapper[4948]: I0220 08:47:48.236262 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhz5" event={"ID":"10cca249-45b2-462c-a348-5881307f2f98","Type":"ContainerDied","Data":"ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137"} Feb 20 08:47:48 crc kubenswrapper[4948]: I0220 08:47:48.236483 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhz5" event={"ID":"10cca249-45b2-462c-a348-5881307f2f98","Type":"ContainerStarted","Data":"2d7ce3d50e15df80a9a9553446d8695357118096db53c5e7f61047f7ee87b249"} Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.116464 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.119365 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.128304 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.128824 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.129420 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.130007 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-tbg4r" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.137656 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.140680 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.140720 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.140857 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.140915 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtx4h\" (UniqueName: \"kubernetes.io/projected/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-kube-api-access-dtx4h\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.141058 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.141088 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-config-data\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.141156 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.141188 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.141210 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.242821 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.242874 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtx4h\" (UniqueName: \"kubernetes.io/projected/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-kube-api-access-dtx4h\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.242955 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.242995 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-config-data\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.243046 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.243074 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.243099 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.243171 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.243192 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.243522 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.243567 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.245839 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-config-data\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.246671 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.250762 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.251085 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.252244 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhz5" event={"ID":"10cca249-45b2-462c-a348-5881307f2f98","Type":"ContainerStarted","Data":"0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1"} Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.253644 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.265622 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtx4h\" (UniqueName: \"kubernetes.io/projected/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-kube-api-access-dtx4h\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.271198 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.281633 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.453854 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 20 08:47:49 crc kubenswrapper[4948]: I0220 08:47:49.916990 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 20 08:47:50 crc kubenswrapper[4948]: I0220 08:47:50.266415 4948 generic.go:334] "Generic (PLEG): container finished" podID="10cca249-45b2-462c-a348-5881307f2f98" containerID="0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1" exitCode=0 Feb 20 08:47:50 crc kubenswrapper[4948]: I0220 08:47:50.266475 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhz5" event={"ID":"10cca249-45b2-462c-a348-5881307f2f98","Type":"ContainerDied","Data":"0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1"} Feb 20 08:47:50 crc kubenswrapper[4948]: I0220 08:47:50.269887 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0fefa0c4-ece9-4caf-8b41-8a64472c53e6","Type":"ContainerStarted","Data":"8d7b11cefa454a0dc17ba4434816118eb89ddebb724fafd4bee4a5317c66243c"} Feb 20 08:47:51 crc kubenswrapper[4948]: I0220 08:47:51.292188 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhz5" event={"ID":"10cca249-45b2-462c-a348-5881307f2f98","Type":"ContainerStarted","Data":"c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84"} Feb 20 08:47:51 crc kubenswrapper[4948]: I0220 08:47:51.315918 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jbhz5" podStartSLOduration=2.900921641 podStartE2EDuration="5.315902015s" podCreationTimestamp="2026-02-20 08:47:46 +0000 UTC" firstStartedPulling="2026-02-20 08:47:48.239431566 +0000 UTC m=+2517.213926436" lastFinishedPulling="2026-02-20 08:47:50.65441198 +0000 UTC m=+2519.628906810" observedRunningTime="2026-02-20 08:47:51.310796568 +0000 UTC m=+2520.285291388" watchObservedRunningTime="2026-02-20 08:47:51.315902015 +0000 UTC m=+2520.290396835" Feb 20 08:47:57 crc kubenswrapper[4948]: I0220 08:47:57.300609 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:57 crc kubenswrapper[4948]: I0220 08:47:57.301218 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:57 crc kubenswrapper[4948]: I0220 08:47:57.376148 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:57 crc kubenswrapper[4948]: I0220 08:47:57.435625 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:47:57 crc kubenswrapper[4948]: I0220 08:47:57.616536 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhz5"] Feb 20 08:47:59 crc kubenswrapper[4948]: I0220 08:47:59.357223 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jbhz5" podUID="10cca249-45b2-462c-a348-5881307f2f98" containerName="registry-server" containerID="cri-o://c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84" gracePeriod=2 Feb 20 08:47:59 crc kubenswrapper[4948]: I0220 08:47:59.886467 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.065652 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7p69\" (UniqueName: \"kubernetes.io/projected/10cca249-45b2-462c-a348-5881307f2f98-kube-api-access-l7p69\") pod \"10cca249-45b2-462c-a348-5881307f2f98\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.065776 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-utilities\") pod \"10cca249-45b2-462c-a348-5881307f2f98\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.065922 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-catalog-content\") pod \"10cca249-45b2-462c-a348-5881307f2f98\" (UID: \"10cca249-45b2-462c-a348-5881307f2f98\") " Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.066569 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-utilities" (OuterVolumeSpecName: "utilities") pod "10cca249-45b2-462c-a348-5881307f2f98" (UID: "10cca249-45b2-462c-a348-5881307f2f98"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.074783 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10cca249-45b2-462c-a348-5881307f2f98-kube-api-access-l7p69" (OuterVolumeSpecName: "kube-api-access-l7p69") pod "10cca249-45b2-462c-a348-5881307f2f98" (UID: "10cca249-45b2-462c-a348-5881307f2f98"). InnerVolumeSpecName "kube-api-access-l7p69". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.087848 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10cca249-45b2-462c-a348-5881307f2f98" (UID: "10cca249-45b2-462c-a348-5881307f2f98"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.167838 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7p69\" (UniqueName: \"kubernetes.io/projected/10cca249-45b2-462c-a348-5881307f2f98-kube-api-access-l7p69\") on node \"crc\" DevicePath \"\"" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.167872 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.167907 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10cca249-45b2-462c-a348-5881307f2f98-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.369106 4948 generic.go:334] "Generic (PLEG): container finished" podID="10cca249-45b2-462c-a348-5881307f2f98" containerID="c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84" exitCode=0 Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.369141 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbhz5" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.369170 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhz5" event={"ID":"10cca249-45b2-462c-a348-5881307f2f98","Type":"ContainerDied","Data":"c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84"} Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.369218 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhz5" event={"ID":"10cca249-45b2-462c-a348-5881307f2f98","Type":"ContainerDied","Data":"2d7ce3d50e15df80a9a9553446d8695357118096db53c5e7f61047f7ee87b249"} Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.369247 4948 scope.go:117] "RemoveContainer" containerID="c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84" Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.419312 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhz5"] Feb 20 08:48:00 crc kubenswrapper[4948]: I0220 08:48:00.426911 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhz5"] Feb 20 08:48:01 crc kubenswrapper[4948]: I0220 08:48:01.738277 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10cca249-45b2-462c-a348-5881307f2f98" path="/var/lib/kubelet/pods/10cca249-45b2-462c-a348-5881307f2f98/volumes" Feb 20 08:48:17 crc kubenswrapper[4948]: I0220 08:48:16.998822 4948 scope.go:117] "RemoveContainer" containerID="0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1" Feb 20 08:48:17 crc kubenswrapper[4948]: I0220 08:48:17.027640 4948 scope.go:117] "RemoveContainer" containerID="ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137" Feb 20 08:48:17 crc kubenswrapper[4948]: I0220 08:48:17.088309 4948 scope.go:117] "RemoveContainer" containerID="c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84" Feb 20 08:48:17 crc kubenswrapper[4948]: E0220 08:48:17.088791 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84\": container with ID starting with c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84 not found: ID does not exist" containerID="c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84" Feb 20 08:48:17 crc kubenswrapper[4948]: I0220 08:48:17.088842 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84"} err="failed to get container status \"c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84\": rpc error: code = NotFound desc = could not find container \"c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84\": container with ID starting with c4c8305352c501549986a399e8dfcfab3c6e25d6963a2311f98753f12586ab84 not found: ID does not exist" Feb 20 08:48:17 crc kubenswrapper[4948]: I0220 08:48:17.088873 4948 scope.go:117] "RemoveContainer" containerID="0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1" Feb 20 08:48:17 crc kubenswrapper[4948]: E0220 08:48:17.089402 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1\": container with ID starting with 0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1 not found: ID does not exist" containerID="0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1" Feb 20 08:48:17 crc kubenswrapper[4948]: I0220 08:48:17.089439 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1"} err="failed to get container status \"0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1\": rpc error: code = NotFound desc = could not find container \"0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1\": container with ID starting with 0e3d35ab9b1e1ccdbcfb45233cef81d9c8c8ee3fc008c103fa0e76b0aabfaac1 not found: ID does not exist" Feb 20 08:48:17 crc kubenswrapper[4948]: I0220 08:48:17.089464 4948 scope.go:117] "RemoveContainer" containerID="ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137" Feb 20 08:48:17 crc kubenswrapper[4948]: E0220 08:48:17.089782 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137\": container with ID starting with ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137 not found: ID does not exist" containerID="ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137" Feb 20 08:48:17 crc kubenswrapper[4948]: I0220 08:48:17.089817 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137"} err="failed to get container status \"ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137\": rpc error: code = NotFound desc = could not find container \"ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137\": container with ID starting with ed45a39cd5cc8d8919b2b3dce997a6b1812e40e55297096de6fbf28e3153a137 not found: ID does not exist" Feb 20 08:48:17 crc kubenswrapper[4948]: E0220 08:48:17.133850 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Feb 20 08:48:17 crc kubenswrapper[4948]: E0220 08:48:17.134067 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dtx4h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(0fefa0c4-ece9-4caf-8b41-8a64472c53e6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 08:48:17 crc kubenswrapper[4948]: E0220 08:48:17.136210 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="0fefa0c4-ece9-4caf-8b41-8a64472c53e6" Feb 20 08:48:17 crc kubenswrapper[4948]: E0220 08:48:17.574559 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="0fefa0c4-ece9-4caf-8b41-8a64472c53e6" Feb 20 08:48:30 crc kubenswrapper[4948]: I0220 08:48:30.151891 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 20 08:48:31 crc kubenswrapper[4948]: I0220 08:48:31.756198 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0fefa0c4-ece9-4caf-8b41-8a64472c53e6","Type":"ContainerStarted","Data":"7c1a082fb78288f20cddeb8b89d13f95d3e809472fa926ab7427ecd860311a72"} Feb 20 08:49:38 crc kubenswrapper[4948]: I0220 08:49:38.025615 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:49:38 crc kubenswrapper[4948]: I0220 08:49:38.026448 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:50:08 crc kubenswrapper[4948]: I0220 08:50:08.025603 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:50:08 crc kubenswrapper[4948]: I0220 08:50:08.027537 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:50:38 crc kubenswrapper[4948]: I0220 08:50:38.025103 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:50:38 crc kubenswrapper[4948]: I0220 08:50:38.025787 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:50:38 crc kubenswrapper[4948]: I0220 08:50:38.025850 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:50:38 crc kubenswrapper[4948]: I0220 08:50:38.026606 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dc74db389f6b091911817b21cddd65e37098545ff6852fc34fb4458a2808d454"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:50:38 crc kubenswrapper[4948]: I0220 08:50:38.026659 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://dc74db389f6b091911817b21cddd65e37098545ff6852fc34fb4458a2808d454" gracePeriod=600 Feb 20 08:50:39 crc kubenswrapper[4948]: I0220 08:50:39.080594 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="dc74db389f6b091911817b21cddd65e37098545ff6852fc34fb4458a2808d454" exitCode=0 Feb 20 08:50:39 crc kubenswrapper[4948]: I0220 08:50:39.080657 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"dc74db389f6b091911817b21cddd65e37098545ff6852fc34fb4458a2808d454"} Feb 20 08:50:39 crc kubenswrapper[4948]: I0220 08:50:39.081254 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5"} Feb 20 08:50:39 crc kubenswrapper[4948]: I0220 08:50:39.081286 4948 scope.go:117] "RemoveContainer" containerID="4fb93478078fda70eb4d21d889559b800c97d003f6db6f528fc347ff8dbb5707" Feb 20 08:50:39 crc kubenswrapper[4948]: I0220 08:50:39.106791 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=130.880422264 podStartE2EDuration="2m51.106776928s" podCreationTimestamp="2026-02-20 08:47:48 +0000 UTC" firstStartedPulling="2026-02-20 08:47:49.92155434 +0000 UTC m=+2518.896049160" lastFinishedPulling="2026-02-20 08:48:30.147908964 +0000 UTC m=+2559.122403824" observedRunningTime="2026-02-20 08:48:31.783936266 +0000 UTC m=+2560.758431126" watchObservedRunningTime="2026-02-20 08:50:39.106776928 +0000 UTC m=+2688.081271748" Feb 20 08:52:38 crc kubenswrapper[4948]: I0220 08:52:38.025018 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:52:38 crc kubenswrapper[4948]: I0220 08:52:38.025517 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:53:08 crc kubenswrapper[4948]: I0220 08:53:08.024760 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:53:08 crc kubenswrapper[4948]: I0220 08:53:08.025413 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.025245 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.026108 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.026191 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.027407 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.027538 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" gracePeriod=600 Feb 20 08:53:38 crc kubenswrapper[4948]: E0220 08:53:38.149581 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.981616 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" exitCode=0 Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.981689 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5"} Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.982159 4948 scope.go:117] "RemoveContainer" containerID="dc74db389f6b091911817b21cddd65e37098545ff6852fc34fb4458a2808d454" Feb 20 08:53:38 crc kubenswrapper[4948]: I0220 08:53:38.983026 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:53:38 crc kubenswrapper[4948]: E0220 08:53:38.983480 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.820921 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dqcs5"] Feb 20 08:53:44 crc kubenswrapper[4948]: E0220 08:53:44.821788 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10cca249-45b2-462c-a348-5881307f2f98" containerName="extract-utilities" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.821802 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="10cca249-45b2-462c-a348-5881307f2f98" containerName="extract-utilities" Feb 20 08:53:44 crc kubenswrapper[4948]: E0220 08:53:44.821820 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10cca249-45b2-462c-a348-5881307f2f98" containerName="registry-server" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.821826 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="10cca249-45b2-462c-a348-5881307f2f98" containerName="registry-server" Feb 20 08:53:44 crc kubenswrapper[4948]: E0220 08:53:44.821845 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10cca249-45b2-462c-a348-5881307f2f98" containerName="extract-content" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.821852 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="10cca249-45b2-462c-a348-5881307f2f98" containerName="extract-content" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.822043 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="10cca249-45b2-462c-a348-5881307f2f98" containerName="registry-server" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.823364 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.833353 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dqcs5"] Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.925194 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-catalog-content\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.925797 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q77f9\" (UniqueName: \"kubernetes.io/projected/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-kube-api-access-q77f9\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:44 crc kubenswrapper[4948]: I0220 08:53:44.925995 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-utilities\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:45 crc kubenswrapper[4948]: I0220 08:53:45.027621 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-catalog-content\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:45 crc kubenswrapper[4948]: I0220 08:53:45.027740 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q77f9\" (UniqueName: \"kubernetes.io/projected/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-kube-api-access-q77f9\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:45 crc kubenswrapper[4948]: I0220 08:53:45.027798 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-utilities\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:45 crc kubenswrapper[4948]: I0220 08:53:45.028247 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-catalog-content\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:45 crc kubenswrapper[4948]: I0220 08:53:45.028277 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-utilities\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:45 crc kubenswrapper[4948]: I0220 08:53:45.053327 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q77f9\" (UniqueName: \"kubernetes.io/projected/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-kube-api-access-q77f9\") pod \"redhat-operators-dqcs5\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:45 crc kubenswrapper[4948]: I0220 08:53:45.174177 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:45 crc kubenswrapper[4948]: I0220 08:53:45.657316 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dqcs5"] Feb 20 08:53:46 crc kubenswrapper[4948]: I0220 08:53:46.045172 4948 generic.go:334] "Generic (PLEG): container finished" podID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerID="e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620" exitCode=0 Feb 20 08:53:46 crc kubenswrapper[4948]: I0220 08:53:46.045443 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqcs5" event={"ID":"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab","Type":"ContainerDied","Data":"e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620"} Feb 20 08:53:46 crc kubenswrapper[4948]: I0220 08:53:46.045522 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqcs5" event={"ID":"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab","Type":"ContainerStarted","Data":"82329664b3d13e9d5f6f72bfbd0c9bfcb4cdedf3613453d9b8308754600e8fce"} Feb 20 08:53:46 crc kubenswrapper[4948]: I0220 08:53:46.047187 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:53:48 crc kubenswrapper[4948]: I0220 08:53:48.066193 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqcs5" event={"ID":"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab","Type":"ContainerStarted","Data":"cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f"} Feb 20 08:53:51 crc kubenswrapper[4948]: I0220 08:53:51.094799 4948 generic.go:334] "Generic (PLEG): container finished" podID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerID="cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f" exitCode=0 Feb 20 08:53:51 crc kubenswrapper[4948]: I0220 08:53:51.094899 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqcs5" event={"ID":"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab","Type":"ContainerDied","Data":"cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f"} Feb 20 08:53:52 crc kubenswrapper[4948]: I0220 08:53:52.107331 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqcs5" event={"ID":"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab","Type":"ContainerStarted","Data":"b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d"} Feb 20 08:53:52 crc kubenswrapper[4948]: I0220 08:53:52.131110 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dqcs5" podStartSLOduration=2.700344155 podStartE2EDuration="8.131081336s" podCreationTimestamp="2026-02-20 08:53:44 +0000 UTC" firstStartedPulling="2026-02-20 08:53:46.046885546 +0000 UTC m=+2875.021380366" lastFinishedPulling="2026-02-20 08:53:51.477622727 +0000 UTC m=+2880.452117547" observedRunningTime="2026-02-20 08:53:52.128712518 +0000 UTC m=+2881.103207338" watchObservedRunningTime="2026-02-20 08:53:52.131081336 +0000 UTC m=+2881.105576186" Feb 20 08:53:52 crc kubenswrapper[4948]: I0220 08:53:52.722797 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:53:52 crc kubenswrapper[4948]: E0220 08:53:52.723173 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:53:55 crc kubenswrapper[4948]: I0220 08:53:55.175052 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:55 crc kubenswrapper[4948]: I0220 08:53:55.175405 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:53:56 crc kubenswrapper[4948]: I0220 08:53:56.280605 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dqcs5" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="registry-server" probeResult="failure" output=< Feb 20 08:53:56 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 08:53:56 crc kubenswrapper[4948]: > Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.230920 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7gx58"] Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.234201 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.246393 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7gx58"] Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.295489 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-catalog-content\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.295573 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtpqf\" (UniqueName: \"kubernetes.io/projected/862b1b6e-55b5-4038-97d3-c1d7bacd7952-kube-api-access-wtpqf\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.295764 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-utilities\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.397811 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-utilities\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.398112 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-catalog-content\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.398155 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtpqf\" (UniqueName: \"kubernetes.io/projected/862b1b6e-55b5-4038-97d3-c1d7bacd7952-kube-api-access-wtpqf\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.398710 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-utilities\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.398748 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-catalog-content\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.419288 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtpqf\" (UniqueName: \"kubernetes.io/projected/862b1b6e-55b5-4038-97d3-c1d7bacd7952-kube-api-access-wtpqf\") pod \"community-operators-7gx58\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:58 crc kubenswrapper[4948]: I0220 08:53:58.566350 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:53:59 crc kubenswrapper[4948]: I0220 08:53:59.032901 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7gx58"] Feb 20 08:53:59 crc kubenswrapper[4948]: I0220 08:53:59.184205 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gx58" event={"ID":"862b1b6e-55b5-4038-97d3-c1d7bacd7952","Type":"ContainerStarted","Data":"57a392d54b60ec1f9ea44e12e9c7761f58441d678cc91dba5253e1e486cc1739"} Feb 20 08:54:00 crc kubenswrapper[4948]: I0220 08:54:00.201198 4948 generic.go:334] "Generic (PLEG): container finished" podID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerID="5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072" exitCode=0 Feb 20 08:54:00 crc kubenswrapper[4948]: I0220 08:54:00.201271 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gx58" event={"ID":"862b1b6e-55b5-4038-97d3-c1d7bacd7952","Type":"ContainerDied","Data":"5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072"} Feb 20 08:54:02 crc kubenswrapper[4948]: I0220 08:54:02.516238 4948 generic.go:334] "Generic (PLEG): container finished" podID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerID="4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6" exitCode=0 Feb 20 08:54:02 crc kubenswrapper[4948]: I0220 08:54:02.516764 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gx58" event={"ID":"862b1b6e-55b5-4038-97d3-c1d7bacd7952","Type":"ContainerDied","Data":"4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6"} Feb 20 08:54:03 crc kubenswrapper[4948]: I0220 08:54:03.530301 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gx58" event={"ID":"862b1b6e-55b5-4038-97d3-c1d7bacd7952","Type":"ContainerStarted","Data":"72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e"} Feb 20 08:54:03 crc kubenswrapper[4948]: I0220 08:54:03.562935 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7gx58" podStartSLOduration=2.8086998359999997 podStartE2EDuration="5.562911118s" podCreationTimestamp="2026-02-20 08:53:58 +0000 UTC" firstStartedPulling="2026-02-20 08:54:00.204362761 +0000 UTC m=+2889.178857581" lastFinishedPulling="2026-02-20 08:54:02.958574013 +0000 UTC m=+2891.933068863" observedRunningTime="2026-02-20 08:54:03.553469047 +0000 UTC m=+2892.527963877" watchObservedRunningTime="2026-02-20 08:54:03.562911118 +0000 UTC m=+2892.537405948" Feb 20 08:54:03 crc kubenswrapper[4948]: I0220 08:54:03.722919 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:54:03 crc kubenswrapper[4948]: E0220 08:54:03.723225 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:54:05 crc kubenswrapper[4948]: I0220 08:54:05.228786 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:54:05 crc kubenswrapper[4948]: I0220 08:54:05.284187 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:54:05 crc kubenswrapper[4948]: I0220 08:54:05.460284 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dqcs5"] Feb 20 08:54:06 crc kubenswrapper[4948]: I0220 08:54:06.566189 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dqcs5" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="registry-server" containerID="cri-o://b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d" gracePeriod=2 Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.097414 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.171619 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q77f9\" (UniqueName: \"kubernetes.io/projected/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-kube-api-access-q77f9\") pod \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.171757 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-catalog-content\") pod \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.172084 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-utilities\") pod \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\" (UID: \"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab\") " Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.172760 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-utilities" (OuterVolumeSpecName: "utilities") pod "5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" (UID: "5cb0bef2-1a21-46e7-905d-6dc6d0c50cab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.172904 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.177794 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-kube-api-access-q77f9" (OuterVolumeSpecName: "kube-api-access-q77f9") pod "5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" (UID: "5cb0bef2-1a21-46e7-905d-6dc6d0c50cab"). InnerVolumeSpecName "kube-api-access-q77f9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.275465 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q77f9\" (UniqueName: \"kubernetes.io/projected/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-kube-api-access-q77f9\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.285916 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" (UID: "5cb0bef2-1a21-46e7-905d-6dc6d0c50cab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.377253 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.576103 4948 generic.go:334] "Generic (PLEG): container finished" podID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerID="b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d" exitCode=0 Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.576147 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqcs5" event={"ID":"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab","Type":"ContainerDied","Data":"b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d"} Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.576169 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqcs5" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.576301 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqcs5" event={"ID":"5cb0bef2-1a21-46e7-905d-6dc6d0c50cab","Type":"ContainerDied","Data":"82329664b3d13e9d5f6f72bfbd0c9bfcb4cdedf3613453d9b8308754600e8fce"} Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.576334 4948 scope.go:117] "RemoveContainer" containerID="b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.616762 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dqcs5"] Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.619252 4948 scope.go:117] "RemoveContainer" containerID="cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.626787 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dqcs5"] Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.646651 4948 scope.go:117] "RemoveContainer" containerID="e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.679028 4948 scope.go:117] "RemoveContainer" containerID="b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d" Feb 20 08:54:07 crc kubenswrapper[4948]: E0220 08:54:07.679488 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d\": container with ID starting with b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d not found: ID does not exist" containerID="b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.679520 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d"} err="failed to get container status \"b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d\": rpc error: code = NotFound desc = could not find container \"b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d\": container with ID starting with b8438f9e17f81afa0b3fb46a37e2e751ee4f1773769441d7d00fc72148764d6d not found: ID does not exist" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.679544 4948 scope.go:117] "RemoveContainer" containerID="cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f" Feb 20 08:54:07 crc kubenswrapper[4948]: E0220 08:54:07.679888 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f\": container with ID starting with cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f not found: ID does not exist" containerID="cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.679941 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f"} err="failed to get container status \"cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f\": rpc error: code = NotFound desc = could not find container \"cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f\": container with ID starting with cdc48d4ad6f6ad6329e706c857c3263754fc1d230af2ba67afcff789d58a0f8f not found: ID does not exist" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.679985 4948 scope.go:117] "RemoveContainer" containerID="e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620" Feb 20 08:54:07 crc kubenswrapper[4948]: E0220 08:54:07.680252 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620\": container with ID starting with e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620 not found: ID does not exist" containerID="e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.680282 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620"} err="failed to get container status \"e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620\": rpc error: code = NotFound desc = could not find container \"e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620\": container with ID starting with e61abe7ff267f9bfeba63a25cfd6a81f5a070c5f25f48e0d6d8f167a049e5620 not found: ID does not exist" Feb 20 08:54:07 crc kubenswrapper[4948]: I0220 08:54:07.731794 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" path="/var/lib/kubelet/pods/5cb0bef2-1a21-46e7-905d-6dc6d0c50cab/volumes" Feb 20 08:54:08 crc kubenswrapper[4948]: I0220 08:54:08.567122 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:54:08 crc kubenswrapper[4948]: I0220 08:54:08.568696 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:54:08 crc kubenswrapper[4948]: I0220 08:54:08.631964 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:54:08 crc kubenswrapper[4948]: I0220 08:54:08.696821 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:54:09 crc kubenswrapper[4948]: I0220 08:54:09.861338 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7gx58"] Feb 20 08:54:10 crc kubenswrapper[4948]: I0220 08:54:10.605698 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7gx58" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerName="registry-server" containerID="cri-o://72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e" gracePeriod=2 Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.118543 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.249344 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-catalog-content\") pod \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.249409 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-utilities\") pod \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.249585 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtpqf\" (UniqueName: \"kubernetes.io/projected/862b1b6e-55b5-4038-97d3-c1d7bacd7952-kube-api-access-wtpqf\") pod \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\" (UID: \"862b1b6e-55b5-4038-97d3-c1d7bacd7952\") " Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.250225 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-utilities" (OuterVolumeSpecName: "utilities") pod "862b1b6e-55b5-4038-97d3-c1d7bacd7952" (UID: "862b1b6e-55b5-4038-97d3-c1d7bacd7952"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.255441 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/862b1b6e-55b5-4038-97d3-c1d7bacd7952-kube-api-access-wtpqf" (OuterVolumeSpecName: "kube-api-access-wtpqf") pod "862b1b6e-55b5-4038-97d3-c1d7bacd7952" (UID: "862b1b6e-55b5-4038-97d3-c1d7bacd7952"). InnerVolumeSpecName "kube-api-access-wtpqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.313366 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "862b1b6e-55b5-4038-97d3-c1d7bacd7952" (UID: "862b1b6e-55b5-4038-97d3-c1d7bacd7952"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.352230 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtpqf\" (UniqueName: \"kubernetes.io/projected/862b1b6e-55b5-4038-97d3-c1d7bacd7952-kube-api-access-wtpqf\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.352269 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.352282 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/862b1b6e-55b5-4038-97d3-c1d7bacd7952-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.616961 4948 generic.go:334] "Generic (PLEG): container finished" podID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerID="72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e" exitCode=0 Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.617043 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gx58" event={"ID":"862b1b6e-55b5-4038-97d3-c1d7bacd7952","Type":"ContainerDied","Data":"72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e"} Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.617109 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gx58" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.617137 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gx58" event={"ID":"862b1b6e-55b5-4038-97d3-c1d7bacd7952","Type":"ContainerDied","Data":"57a392d54b60ec1f9ea44e12e9c7761f58441d678cc91dba5253e1e486cc1739"} Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.617177 4948 scope.go:117] "RemoveContainer" containerID="72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.641296 4948 scope.go:117] "RemoveContainer" containerID="4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.663851 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7gx58"] Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.678951 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7gx58"] Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.681207 4948 scope.go:117] "RemoveContainer" containerID="5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.721015 4948 scope.go:117] "RemoveContainer" containerID="72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e" Feb 20 08:54:11 crc kubenswrapper[4948]: E0220 08:54:11.724266 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e\": container with ID starting with 72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e not found: ID does not exist" containerID="72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.724312 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e"} err="failed to get container status \"72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e\": rpc error: code = NotFound desc = could not find container \"72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e\": container with ID starting with 72e21e1986add64697efd8d697ba5a97302e08211fa86c5c48d50e610852427e not found: ID does not exist" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.724338 4948 scope.go:117] "RemoveContainer" containerID="4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6" Feb 20 08:54:11 crc kubenswrapper[4948]: E0220 08:54:11.724788 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6\": container with ID starting with 4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6 not found: ID does not exist" containerID="4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.724817 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6"} err="failed to get container status \"4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6\": rpc error: code = NotFound desc = could not find container \"4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6\": container with ID starting with 4d29bdce5261c62a695692c37c480bd775f65ba1c115d881b2d004a4ee0fd7b6 not found: ID does not exist" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.724832 4948 scope.go:117] "RemoveContainer" containerID="5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072" Feb 20 08:54:11 crc kubenswrapper[4948]: E0220 08:54:11.725155 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072\": container with ID starting with 5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072 not found: ID does not exist" containerID="5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.725187 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072"} err="failed to get container status \"5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072\": rpc error: code = NotFound desc = could not find container \"5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072\": container with ID starting with 5cc46dc46f6bf2327c20993dbc63a5a80831fa33a8858c4268efa9310d184072 not found: ID does not exist" Feb 20 08:54:11 crc kubenswrapper[4948]: I0220 08:54:11.733896 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" path="/var/lib/kubelet/pods/862b1b6e-55b5-4038-97d3-c1d7bacd7952/volumes" Feb 20 08:54:17 crc kubenswrapper[4948]: I0220 08:54:17.723009 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:54:17 crc kubenswrapper[4948]: E0220 08:54:17.724073 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:54:29 crc kubenswrapper[4948]: I0220 08:54:29.723164 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:54:29 crc kubenswrapper[4948]: E0220 08:54:29.724128 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.183543 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6gxcp"] Feb 20 08:54:36 crc kubenswrapper[4948]: E0220 08:54:36.184639 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerName="extract-utilities" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.184660 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerName="extract-utilities" Feb 20 08:54:36 crc kubenswrapper[4948]: E0220 08:54:36.184673 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="extract-content" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.184681 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="extract-content" Feb 20 08:54:36 crc kubenswrapper[4948]: E0220 08:54:36.184699 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="registry-server" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.184707 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="registry-server" Feb 20 08:54:36 crc kubenswrapper[4948]: E0220 08:54:36.184720 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerName="extract-content" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.184727 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerName="extract-content" Feb 20 08:54:36 crc kubenswrapper[4948]: E0220 08:54:36.184756 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="extract-utilities" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.184763 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="extract-utilities" Feb 20 08:54:36 crc kubenswrapper[4948]: E0220 08:54:36.184774 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerName="registry-server" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.184780 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerName="registry-server" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.185026 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="862b1b6e-55b5-4038-97d3-c1d7bacd7952" containerName="registry-server" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.185054 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cb0bef2-1a21-46e7-905d-6dc6d0c50cab" containerName="registry-server" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.186691 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.210021 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6gxcp"] Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.247061 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddsfc\" (UniqueName: \"kubernetes.io/projected/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-kube-api-access-ddsfc\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.248132 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-utilities\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.248234 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-catalog-content\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.350145 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddsfc\" (UniqueName: \"kubernetes.io/projected/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-kube-api-access-ddsfc\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.350239 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-utilities\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.350289 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-catalog-content\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.351006 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-catalog-content\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.351029 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-utilities\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.371884 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddsfc\" (UniqueName: \"kubernetes.io/projected/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-kube-api-access-ddsfc\") pod \"certified-operators-6gxcp\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:36 crc kubenswrapper[4948]: I0220 08:54:36.515659 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:37 crc kubenswrapper[4948]: I0220 08:54:37.083095 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6gxcp"] Feb 20 08:54:37 crc kubenswrapper[4948]: I0220 08:54:37.887048 4948 generic.go:334] "Generic (PLEG): container finished" podID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerID="d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73" exitCode=0 Feb 20 08:54:37 crc kubenswrapper[4948]: I0220 08:54:37.887115 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gxcp" event={"ID":"0f0a4e55-97e8-4b22-8299-ea443d51a3c1","Type":"ContainerDied","Data":"d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73"} Feb 20 08:54:37 crc kubenswrapper[4948]: I0220 08:54:37.887181 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gxcp" event={"ID":"0f0a4e55-97e8-4b22-8299-ea443d51a3c1","Type":"ContainerStarted","Data":"653f289bbb6954dacf1dccca73b66f05c7ef3ea91b0fb94146d797c22d4827b2"} Feb 20 08:54:38 crc kubenswrapper[4948]: I0220 08:54:38.899375 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gxcp" event={"ID":"0f0a4e55-97e8-4b22-8299-ea443d51a3c1","Type":"ContainerStarted","Data":"1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3"} Feb 20 08:54:39 crc kubenswrapper[4948]: I0220 08:54:39.911966 4948 generic.go:334] "Generic (PLEG): container finished" podID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerID="1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3" exitCode=0 Feb 20 08:54:39 crc kubenswrapper[4948]: I0220 08:54:39.912032 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gxcp" event={"ID":"0f0a4e55-97e8-4b22-8299-ea443d51a3c1","Type":"ContainerDied","Data":"1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3"} Feb 20 08:54:40 crc kubenswrapper[4948]: I0220 08:54:40.723096 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:54:40 crc kubenswrapper[4948]: E0220 08:54:40.723902 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:54:40 crc kubenswrapper[4948]: I0220 08:54:40.925475 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gxcp" event={"ID":"0f0a4e55-97e8-4b22-8299-ea443d51a3c1","Type":"ContainerStarted","Data":"1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34"} Feb 20 08:54:40 crc kubenswrapper[4948]: I0220 08:54:40.949672 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6gxcp" podStartSLOduration=2.526167647 podStartE2EDuration="4.949644825s" podCreationTimestamp="2026-02-20 08:54:36 +0000 UTC" firstStartedPulling="2026-02-20 08:54:37.889425737 +0000 UTC m=+2926.863920557" lastFinishedPulling="2026-02-20 08:54:40.312902915 +0000 UTC m=+2929.287397735" observedRunningTime="2026-02-20 08:54:40.943043453 +0000 UTC m=+2929.917538273" watchObservedRunningTime="2026-02-20 08:54:40.949644825 +0000 UTC m=+2929.924139645" Feb 20 08:54:46 crc kubenswrapper[4948]: I0220 08:54:46.517037 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:46 crc kubenswrapper[4948]: I0220 08:54:46.517700 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:46 crc kubenswrapper[4948]: I0220 08:54:46.558112 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:47 crc kubenswrapper[4948]: I0220 08:54:47.042722 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:47 crc kubenswrapper[4948]: I0220 08:54:47.105337 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6gxcp"] Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.004525 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6gxcp" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerName="registry-server" containerID="cri-o://1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34" gracePeriod=2 Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.571205 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.639290 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-catalog-content\") pod \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.639511 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddsfc\" (UniqueName: \"kubernetes.io/projected/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-kube-api-access-ddsfc\") pod \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.639628 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-utilities\") pod \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\" (UID: \"0f0a4e55-97e8-4b22-8299-ea443d51a3c1\") " Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.642055 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-utilities" (OuterVolumeSpecName: "utilities") pod "0f0a4e55-97e8-4b22-8299-ea443d51a3c1" (UID: "0f0a4e55-97e8-4b22-8299-ea443d51a3c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.652187 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-kube-api-access-ddsfc" (OuterVolumeSpecName: "kube-api-access-ddsfc") pod "0f0a4e55-97e8-4b22-8299-ea443d51a3c1" (UID: "0f0a4e55-97e8-4b22-8299-ea443d51a3c1"). InnerVolumeSpecName "kube-api-access-ddsfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.748462 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddsfc\" (UniqueName: \"kubernetes.io/projected/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-kube-api-access-ddsfc\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.748815 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.795112 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f0a4e55-97e8-4b22-8299-ea443d51a3c1" (UID: "0f0a4e55-97e8-4b22-8299-ea443d51a3c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:54:49 crc kubenswrapper[4948]: I0220 08:54:49.851617 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f0a4e55-97e8-4b22-8299-ea443d51a3c1-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.016998 4948 generic.go:334] "Generic (PLEG): container finished" podID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerID="1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34" exitCode=0 Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.017058 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gxcp" event={"ID":"0f0a4e55-97e8-4b22-8299-ea443d51a3c1","Type":"ContainerDied","Data":"1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34"} Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.017095 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gxcp" event={"ID":"0f0a4e55-97e8-4b22-8299-ea443d51a3c1","Type":"ContainerDied","Data":"653f289bbb6954dacf1dccca73b66f05c7ef3ea91b0fb94146d797c22d4827b2"} Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.017117 4948 scope.go:117] "RemoveContainer" containerID="1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.017296 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gxcp" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.040679 4948 scope.go:117] "RemoveContainer" containerID="1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.083169 4948 scope.go:117] "RemoveContainer" containerID="d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.085953 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6gxcp"] Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.097402 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6gxcp"] Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.133836 4948 scope.go:117] "RemoveContainer" containerID="1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34" Feb 20 08:54:50 crc kubenswrapper[4948]: E0220 08:54:50.134449 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34\": container with ID starting with 1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34 not found: ID does not exist" containerID="1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.134482 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34"} err="failed to get container status \"1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34\": rpc error: code = NotFound desc = could not find container \"1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34\": container with ID starting with 1ae5da12074875624d809ff770ce45f4eed832a8a73b4e198a873dfd6f9b6b34 not found: ID does not exist" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.134507 4948 scope.go:117] "RemoveContainer" containerID="1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3" Feb 20 08:54:50 crc kubenswrapper[4948]: E0220 08:54:50.135556 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3\": container with ID starting with 1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3 not found: ID does not exist" containerID="1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.135593 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3"} err="failed to get container status \"1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3\": rpc error: code = NotFound desc = could not find container \"1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3\": container with ID starting with 1028c59d0521f3ece4f33a9c808df88e3f3aeb440639acaf3addacfc8e3fc9f3 not found: ID does not exist" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.135612 4948 scope.go:117] "RemoveContainer" containerID="d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73" Feb 20 08:54:50 crc kubenswrapper[4948]: E0220 08:54:50.135871 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73\": container with ID starting with d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73 not found: ID does not exist" containerID="d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73" Feb 20 08:54:50 crc kubenswrapper[4948]: I0220 08:54:50.135897 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73"} err="failed to get container status \"d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73\": rpc error: code = NotFound desc = could not find container \"d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73\": container with ID starting with d0716005a1627047c505cf076f1e700308f61ddebea1580bef3289c1f5855a73 not found: ID does not exist" Feb 20 08:54:51 crc kubenswrapper[4948]: I0220 08:54:51.741830 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" path="/var/lib/kubelet/pods/0f0a4e55-97e8-4b22-8299-ea443d51a3c1/volumes" Feb 20 08:54:54 crc kubenswrapper[4948]: I0220 08:54:54.723153 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:54:54 crc kubenswrapper[4948]: E0220 08:54:54.724166 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:55:09 crc kubenswrapper[4948]: I0220 08:55:09.723150 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:55:09 crc kubenswrapper[4948]: E0220 08:55:09.724016 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:55:20 crc kubenswrapper[4948]: I0220 08:55:20.723367 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:55:20 crc kubenswrapper[4948]: E0220 08:55:20.724796 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:55:33 crc kubenswrapper[4948]: I0220 08:55:33.724643 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:55:33 crc kubenswrapper[4948]: E0220 08:55:33.725894 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:55:47 crc kubenswrapper[4948]: I0220 08:55:47.747193 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:55:47 crc kubenswrapper[4948]: E0220 08:55:47.749372 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:56:00 crc kubenswrapper[4948]: I0220 08:56:00.722957 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:56:00 crc kubenswrapper[4948]: E0220 08:56:00.723749 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:56:11 crc kubenswrapper[4948]: I0220 08:56:11.728796 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:56:11 crc kubenswrapper[4948]: E0220 08:56:11.730052 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:56:26 crc kubenswrapper[4948]: I0220 08:56:26.724214 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:56:26 crc kubenswrapper[4948]: E0220 08:56:26.725765 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:56:41 crc kubenswrapper[4948]: I0220 08:56:41.736878 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:56:41 crc kubenswrapper[4948]: E0220 08:56:41.738370 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:56:53 crc kubenswrapper[4948]: I0220 08:56:53.722749 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:56:53 crc kubenswrapper[4948]: E0220 08:56:53.724004 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:57:05 crc kubenswrapper[4948]: I0220 08:57:05.723232 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:57:05 crc kubenswrapper[4948]: E0220 08:57:05.724391 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:57:17 crc kubenswrapper[4948]: I0220 08:57:17.723495 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:57:17 crc kubenswrapper[4948]: E0220 08:57:17.727215 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:57:31 crc kubenswrapper[4948]: I0220 08:57:31.729696 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:57:31 crc kubenswrapper[4948]: E0220 08:57:31.730486 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:57:42 crc kubenswrapper[4948]: I0220 08:57:42.722312 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:57:42 crc kubenswrapper[4948]: E0220 08:57:42.723112 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:57:56 crc kubenswrapper[4948]: I0220 08:57:56.722942 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:57:56 crc kubenswrapper[4948]: E0220 08:57:56.723828 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:58:10 crc kubenswrapper[4948]: I0220 08:58:10.723145 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:58:10 crc kubenswrapper[4948]: E0220 08:58:10.724575 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:58:24 crc kubenswrapper[4948]: I0220 08:58:24.722239 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:58:24 crc kubenswrapper[4948]: E0220 08:58:24.723108 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:58:36 crc kubenswrapper[4948]: I0220 08:58:36.723154 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:58:36 crc kubenswrapper[4948]: E0220 08:58:36.725801 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 08:58:49 crc kubenswrapper[4948]: I0220 08:58:49.723754 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 08:58:50 crc kubenswrapper[4948]: I0220 08:58:50.521317 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"b0e4da4a6a513a5c47c1f0dd87dcf0b56d3fede7e9008d839ac5623f61075ff8"} Feb 20 08:59:07 crc kubenswrapper[4948]: I0220 08:59:07.759124 4948 generic.go:334] "Generic (PLEG): container finished" podID="0fefa0c4-ece9-4caf-8b41-8a64472c53e6" containerID="7c1a082fb78288f20cddeb8b89d13f95d3e809472fa926ab7427ecd860311a72" exitCode=0 Feb 20 08:59:07 crc kubenswrapper[4948]: I0220 08:59:07.759268 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0fefa0c4-ece9-4caf-8b41-8a64472c53e6","Type":"ContainerDied","Data":"7c1a082fb78288f20cddeb8b89d13f95d3e809472fa926ab7427ecd860311a72"} Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.245400 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.361597 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.361707 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.361780 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config-secret\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.361881 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtx4h\" (UniqueName: \"kubernetes.io/projected/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-kube-api-access-dtx4h\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.361946 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-config-data\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.362063 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ca-certs\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.362210 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ssh-key\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.362450 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-temporary\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.362543 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-workdir\") pod \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\" (UID: \"0fefa0c4-ece9-4caf-8b41-8a64472c53e6\") " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.362905 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.362706 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-config-data" (OuterVolumeSpecName: "config-data") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.363508 4948 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.363551 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.367420 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.368007 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "test-operator-logs") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.368137 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-kube-api-access-dtx4h" (OuterVolumeSpecName: "kube-api-access-dtx4h") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "kube-api-access-dtx4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.393336 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.396057 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.402339 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.414556 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "0fefa0c4-ece9-4caf-8b41-8a64472c53e6" (UID: "0fefa0c4-ece9-4caf-8b41-8a64472c53e6"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.465798 4948 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.465855 4948 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.465876 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtx4h\" (UniqueName: \"kubernetes.io/projected/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-kube-api-access-dtx4h\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.465894 4948 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ca-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.465910 4948 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-ssh-key\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.465941 4948 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0fefa0c4-ece9-4caf-8b41-8a64472c53e6-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.466075 4948 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.512043 4948 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.567879 4948 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.785626 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"0fefa0c4-ece9-4caf-8b41-8a64472c53e6","Type":"ContainerDied","Data":"8d7b11cefa454a0dc17ba4434816118eb89ddebb724fafd4bee4a5317c66243c"} Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.785671 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d7b11cefa454a0dc17ba4434816118eb89ddebb724fafd4bee4a5317c66243c" Feb 20 08:59:09 crc kubenswrapper[4948]: I0220 08:59:09.785771 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.378250 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g72b5"] Feb 20 08:59:11 crc kubenswrapper[4948]: E0220 08:59:11.379069 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerName="extract-utilities" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.379086 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerName="extract-utilities" Feb 20 08:59:11 crc kubenswrapper[4948]: E0220 08:59:11.379112 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerName="extract-content" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.379121 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerName="extract-content" Feb 20 08:59:11 crc kubenswrapper[4948]: E0220 08:59:11.379145 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerName="registry-server" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.379153 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerName="registry-server" Feb 20 08:59:11 crc kubenswrapper[4948]: E0220 08:59:11.379177 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fefa0c4-ece9-4caf-8b41-8a64472c53e6" containerName="tempest-tests-tempest-tests-runner" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.379186 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fefa0c4-ece9-4caf-8b41-8a64472c53e6" containerName="tempest-tests-tempest-tests-runner" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.379458 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f0a4e55-97e8-4b22-8299-ea443d51a3c1" containerName="registry-server" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.379477 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fefa0c4-ece9-4caf-8b41-8a64472c53e6" containerName="tempest-tests-tempest-tests-runner" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.395120 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.398224 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g72b5"] Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.523090 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-utilities\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.523149 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srqhw\" (UniqueName: \"kubernetes.io/projected/4c730f22-7ac4-4c78-b897-5aed76dbbd79-kube-api-access-srqhw\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.523220 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-catalog-content\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.624753 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-utilities\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.624800 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srqhw\" (UniqueName: \"kubernetes.io/projected/4c730f22-7ac4-4c78-b897-5aed76dbbd79-kube-api-access-srqhw\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.624839 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-catalog-content\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.625512 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-catalog-content\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.625512 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-utilities\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.643998 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srqhw\" (UniqueName: \"kubernetes.io/projected/4c730f22-7ac4-4c78-b897-5aed76dbbd79-kube-api-access-srqhw\") pod \"redhat-marketplace-g72b5\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:11 crc kubenswrapper[4948]: I0220 08:59:11.726255 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:12 crc kubenswrapper[4948]: I0220 08:59:12.242689 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g72b5"] Feb 20 08:59:12 crc kubenswrapper[4948]: W0220 08:59:12.252361 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c730f22_7ac4_4c78_b897_5aed76dbbd79.slice/crio-834d32f4c058854127072ff368549105e1f01138af7f056fc671dc63c4a176c7 WatchSource:0}: Error finding container 834d32f4c058854127072ff368549105e1f01138af7f056fc671dc63c4a176c7: Status 404 returned error can't find the container with id 834d32f4c058854127072ff368549105e1f01138af7f056fc671dc63c4a176c7 Feb 20 08:59:12 crc kubenswrapper[4948]: I0220 08:59:12.822889 4948 generic.go:334] "Generic (PLEG): container finished" podID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerID="f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821" exitCode=0 Feb 20 08:59:12 crc kubenswrapper[4948]: I0220 08:59:12.822952 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g72b5" event={"ID":"4c730f22-7ac4-4c78-b897-5aed76dbbd79","Type":"ContainerDied","Data":"f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821"} Feb 20 08:59:12 crc kubenswrapper[4948]: I0220 08:59:12.823027 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g72b5" event={"ID":"4c730f22-7ac4-4c78-b897-5aed76dbbd79","Type":"ContainerStarted","Data":"834d32f4c058854127072ff368549105e1f01138af7f056fc671dc63c4a176c7"} Feb 20 08:59:12 crc kubenswrapper[4948]: I0220 08:59:12.824832 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:59:13 crc kubenswrapper[4948]: I0220 08:59:13.836798 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g72b5" event={"ID":"4c730f22-7ac4-4c78-b897-5aed76dbbd79","Type":"ContainerStarted","Data":"f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0"} Feb 20 08:59:13 crc kubenswrapper[4948]: I0220 08:59:13.958444 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 20 08:59:13 crc kubenswrapper[4948]: I0220 08:59:13.960082 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:13 crc kubenswrapper[4948]: I0220 08:59:13.962647 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-tbg4r" Feb 20 08:59:13 crc kubenswrapper[4948]: I0220 08:59:13.981186 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.070252 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"de083766-efe0-42cd-95c5-b87bbeb33b0a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.070564 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9drzn\" (UniqueName: \"kubernetes.io/projected/de083766-efe0-42cd-95c5-b87bbeb33b0a-kube-api-access-9drzn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"de083766-efe0-42cd-95c5-b87bbeb33b0a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.172332 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"de083766-efe0-42cd-95c5-b87bbeb33b0a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.172693 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9drzn\" (UniqueName: \"kubernetes.io/projected/de083766-efe0-42cd-95c5-b87bbeb33b0a-kube-api-access-9drzn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"de083766-efe0-42cd-95c5-b87bbeb33b0a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.172934 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"de083766-efe0-42cd-95c5-b87bbeb33b0a\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.200998 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9drzn\" (UniqueName: \"kubernetes.io/projected/de083766-efe0-42cd-95c5-b87bbeb33b0a-kube-api-access-9drzn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"de083766-efe0-42cd-95c5-b87bbeb33b0a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.219313 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"de083766-efe0-42cd-95c5-b87bbeb33b0a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.289141 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.820308 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.846477 4948 generic.go:334] "Generic (PLEG): container finished" podID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerID="f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0" exitCode=0 Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.847450 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g72b5" event={"ID":"4c730f22-7ac4-4c78-b897-5aed76dbbd79","Type":"ContainerDied","Data":"f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0"} Feb 20 08:59:14 crc kubenswrapper[4948]: I0220 08:59:14.852003 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"de083766-efe0-42cd-95c5-b87bbeb33b0a","Type":"ContainerStarted","Data":"660b8d90d795d12fd9b82a49aa9566597ff1b439beab84abbc27859e9de18cb7"} Feb 20 08:59:16 crc kubenswrapper[4948]: I0220 08:59:16.868881 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g72b5" event={"ID":"4c730f22-7ac4-4c78-b897-5aed76dbbd79","Type":"ContainerStarted","Data":"144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be"} Feb 20 08:59:16 crc kubenswrapper[4948]: I0220 08:59:16.906176 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g72b5" podStartSLOduration=2.804067953 podStartE2EDuration="5.906155814s" podCreationTimestamp="2026-02-20 08:59:11 +0000 UTC" firstStartedPulling="2026-02-20 08:59:12.824563728 +0000 UTC m=+3201.799058548" lastFinishedPulling="2026-02-20 08:59:15.926651559 +0000 UTC m=+3204.901146409" observedRunningTime="2026-02-20 08:59:16.891947213 +0000 UTC m=+3205.866442033" watchObservedRunningTime="2026-02-20 08:59:16.906155814 +0000 UTC m=+3205.880650634" Feb 20 08:59:17 crc kubenswrapper[4948]: I0220 08:59:17.881627 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"de083766-efe0-42cd-95c5-b87bbeb33b0a","Type":"ContainerStarted","Data":"4b8d29446ac83847d90e3bb8e9583f0225f10b403cd18e4c1734d1f7b374eff9"} Feb 20 08:59:17 crc kubenswrapper[4948]: I0220 08:59:17.905304 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=3.044024304 podStartE2EDuration="4.905269482s" podCreationTimestamp="2026-02-20 08:59:13 +0000 UTC" firstStartedPulling="2026-02-20 08:59:14.81243052 +0000 UTC m=+3203.786925350" lastFinishedPulling="2026-02-20 08:59:16.673675688 +0000 UTC m=+3205.648170528" observedRunningTime="2026-02-20 08:59:17.903003766 +0000 UTC m=+3206.877498626" watchObservedRunningTime="2026-02-20 08:59:17.905269482 +0000 UTC m=+3206.879764342" Feb 20 08:59:21 crc kubenswrapper[4948]: I0220 08:59:21.742262 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:21 crc kubenswrapper[4948]: I0220 08:59:21.742818 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:21 crc kubenswrapper[4948]: I0220 08:59:21.792772 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:21 crc kubenswrapper[4948]: I0220 08:59:21.973786 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:22 crc kubenswrapper[4948]: I0220 08:59:22.048277 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g72b5"] Feb 20 08:59:23 crc kubenswrapper[4948]: I0220 08:59:23.949654 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-g72b5" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerName="registry-server" containerID="cri-o://144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be" gracePeriod=2 Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.453721 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.633850 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-catalog-content\") pod \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.634146 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srqhw\" (UniqueName: \"kubernetes.io/projected/4c730f22-7ac4-4c78-b897-5aed76dbbd79-kube-api-access-srqhw\") pod \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.634198 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-utilities\") pod \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\" (UID: \"4c730f22-7ac4-4c78-b897-5aed76dbbd79\") " Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.635230 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-utilities" (OuterVolumeSpecName: "utilities") pod "4c730f22-7ac4-4c78-b897-5aed76dbbd79" (UID: "4c730f22-7ac4-4c78-b897-5aed76dbbd79"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.642757 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c730f22-7ac4-4c78-b897-5aed76dbbd79-kube-api-access-srqhw" (OuterVolumeSpecName: "kube-api-access-srqhw") pod "4c730f22-7ac4-4c78-b897-5aed76dbbd79" (UID: "4c730f22-7ac4-4c78-b897-5aed76dbbd79"). InnerVolumeSpecName "kube-api-access-srqhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.672509 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c730f22-7ac4-4c78-b897-5aed76dbbd79" (UID: "4c730f22-7ac4-4c78-b897-5aed76dbbd79"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.736274 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srqhw\" (UniqueName: \"kubernetes.io/projected/4c730f22-7ac4-4c78-b897-5aed76dbbd79-kube-api-access-srqhw\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.736607 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.736846 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c730f22-7ac4-4c78-b897-5aed76dbbd79-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.965040 4948 generic.go:334] "Generic (PLEG): container finished" podID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerID="144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be" exitCode=0 Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.965117 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g72b5" event={"ID":"4c730f22-7ac4-4c78-b897-5aed76dbbd79","Type":"ContainerDied","Data":"144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be"} Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.965200 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g72b5" event={"ID":"4c730f22-7ac4-4c78-b897-5aed76dbbd79","Type":"ContainerDied","Data":"834d32f4c058854127072ff368549105e1f01138af7f056fc671dc63c4a176c7"} Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.965237 4948 scope.go:117] "RemoveContainer" containerID="144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be" Feb 20 08:59:24 crc kubenswrapper[4948]: I0220 08:59:24.967067 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g72b5" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.001059 4948 scope.go:117] "RemoveContainer" containerID="f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.035547 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g72b5"] Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.043457 4948 scope.go:117] "RemoveContainer" containerID="f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.057579 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-g72b5"] Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.081367 4948 scope.go:117] "RemoveContainer" containerID="144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be" Feb 20 08:59:25 crc kubenswrapper[4948]: E0220 08:59:25.081940 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be\": container with ID starting with 144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be not found: ID does not exist" containerID="144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.082137 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be"} err="failed to get container status \"144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be\": rpc error: code = NotFound desc = could not find container \"144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be\": container with ID starting with 144c046a74ee5454148fd8c9d93123fe6564ef046b5f98222707f50a14a587be not found: ID does not exist" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.082292 4948 scope.go:117] "RemoveContainer" containerID="f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0" Feb 20 08:59:25 crc kubenswrapper[4948]: E0220 08:59:25.082782 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0\": container with ID starting with f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0 not found: ID does not exist" containerID="f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.082930 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0"} err="failed to get container status \"f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0\": rpc error: code = NotFound desc = could not find container \"f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0\": container with ID starting with f990f9b78c70c0197487e4632df495d206960a5cc7ff502f7b84e44e7c850bf0 not found: ID does not exist" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.083107 4948 scope.go:117] "RemoveContainer" containerID="f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821" Feb 20 08:59:25 crc kubenswrapper[4948]: E0220 08:59:25.083611 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821\": container with ID starting with f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821 not found: ID does not exist" containerID="f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.083638 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821"} err="failed to get container status \"f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821\": rpc error: code = NotFound desc = could not find container \"f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821\": container with ID starting with f4c2df2bcbe2ce8c1b34e4d9226525be076033ebb476794a5b25b1871ed87821 not found: ID does not exist" Feb 20 08:59:25 crc kubenswrapper[4948]: I0220 08:59:25.739751 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" path="/var/lib/kubelet/pods/4c730f22-7ac4-4c78-b897-5aed76dbbd79/volumes" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.875339 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8sqsd/must-gather-xzw72"] Feb 20 08:59:37 crc kubenswrapper[4948]: E0220 08:59:37.876377 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerName="extract-utilities" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.876395 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerName="extract-utilities" Feb 20 08:59:37 crc kubenswrapper[4948]: E0220 08:59:37.876419 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerName="registry-server" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.876427 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerName="registry-server" Feb 20 08:59:37 crc kubenswrapper[4948]: E0220 08:59:37.876458 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerName="extract-content" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.876466 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerName="extract-content" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.876720 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c730f22-7ac4-4c78-b897-5aed76dbbd79" containerName="registry-server" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.882376 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.885028 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-8sqsd"/"openshift-service-ca.crt" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.885310 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-8sqsd"/"kube-root-ca.crt" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.898738 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-8sqsd/must-gather-xzw72"] Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.999020 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b25c36d5-5de9-4c98-acf4-495d492def68-must-gather-output\") pod \"must-gather-xzw72\" (UID: \"b25c36d5-5de9-4c98-acf4-495d492def68\") " pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 08:59:37 crc kubenswrapper[4948]: I0220 08:59:37.999241 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsr8v\" (UniqueName: \"kubernetes.io/projected/b25c36d5-5de9-4c98-acf4-495d492def68-kube-api-access-vsr8v\") pod \"must-gather-xzw72\" (UID: \"b25c36d5-5de9-4c98-acf4-495d492def68\") " pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 08:59:38 crc kubenswrapper[4948]: I0220 08:59:38.101526 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b25c36d5-5de9-4c98-acf4-495d492def68-must-gather-output\") pod \"must-gather-xzw72\" (UID: \"b25c36d5-5de9-4c98-acf4-495d492def68\") " pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 08:59:38 crc kubenswrapper[4948]: I0220 08:59:38.101593 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsr8v\" (UniqueName: \"kubernetes.io/projected/b25c36d5-5de9-4c98-acf4-495d492def68-kube-api-access-vsr8v\") pod \"must-gather-xzw72\" (UID: \"b25c36d5-5de9-4c98-acf4-495d492def68\") " pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 08:59:38 crc kubenswrapper[4948]: I0220 08:59:38.102063 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b25c36d5-5de9-4c98-acf4-495d492def68-must-gather-output\") pod \"must-gather-xzw72\" (UID: \"b25c36d5-5de9-4c98-acf4-495d492def68\") " pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 08:59:38 crc kubenswrapper[4948]: I0220 08:59:38.128985 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsr8v\" (UniqueName: \"kubernetes.io/projected/b25c36d5-5de9-4c98-acf4-495d492def68-kube-api-access-vsr8v\") pod \"must-gather-xzw72\" (UID: \"b25c36d5-5de9-4c98-acf4-495d492def68\") " pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 08:59:38 crc kubenswrapper[4948]: I0220 08:59:38.203168 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 08:59:38 crc kubenswrapper[4948]: I0220 08:59:38.671399 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-8sqsd/must-gather-xzw72"] Feb 20 08:59:39 crc kubenswrapper[4948]: I0220 08:59:39.109136 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/must-gather-xzw72" event={"ID":"b25c36d5-5de9-4c98-acf4-495d492def68","Type":"ContainerStarted","Data":"8c112f2a53188eebfa660b390f650533de90ac513fc79f81c317095abbfa670e"} Feb 20 08:59:45 crc kubenswrapper[4948]: I0220 08:59:45.180533 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/must-gather-xzw72" event={"ID":"b25c36d5-5de9-4c98-acf4-495d492def68","Type":"ContainerStarted","Data":"ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4"} Feb 20 08:59:45 crc kubenswrapper[4948]: I0220 08:59:45.181059 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/must-gather-xzw72" event={"ID":"b25c36d5-5de9-4c98-acf4-495d492def68","Type":"ContainerStarted","Data":"4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a"} Feb 20 08:59:45 crc kubenswrapper[4948]: I0220 08:59:45.199604 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-8sqsd/must-gather-xzw72" podStartSLOduration=2.369945358 podStartE2EDuration="8.199584979s" podCreationTimestamp="2026-02-20 08:59:37 +0000 UTC" firstStartedPulling="2026-02-20 08:59:38.684707433 +0000 UTC m=+3227.659202253" lastFinishedPulling="2026-02-20 08:59:44.514347044 +0000 UTC m=+3233.488841874" observedRunningTime="2026-02-20 08:59:45.19556648 +0000 UTC m=+3234.170061300" watchObservedRunningTime="2026-02-20 08:59:45.199584979 +0000 UTC m=+3234.174079819" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.351644 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-2rr8w"] Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.355613 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.359954 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-8sqsd"/"default-dockercfg-mzzkw" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.522151 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp4zr\" (UniqueName: \"kubernetes.io/projected/e1eaa8b0-3f84-4312-b7ac-4cc536093646-kube-api-access-bp4zr\") pod \"crc-debug-2rr8w\" (UID: \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\") " pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.522499 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e1eaa8b0-3f84-4312-b7ac-4cc536093646-host\") pod \"crc-debug-2rr8w\" (UID: \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\") " pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.623726 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp4zr\" (UniqueName: \"kubernetes.io/projected/e1eaa8b0-3f84-4312-b7ac-4cc536093646-kube-api-access-bp4zr\") pod \"crc-debug-2rr8w\" (UID: \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\") " pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.623882 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e1eaa8b0-3f84-4312-b7ac-4cc536093646-host\") pod \"crc-debug-2rr8w\" (UID: \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\") " pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.623990 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e1eaa8b0-3f84-4312-b7ac-4cc536093646-host\") pod \"crc-debug-2rr8w\" (UID: \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\") " pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.641726 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp4zr\" (UniqueName: \"kubernetes.io/projected/e1eaa8b0-3f84-4312-b7ac-4cc536093646-kube-api-access-bp4zr\") pod \"crc-debug-2rr8w\" (UID: \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\") " pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 08:59:49 crc kubenswrapper[4948]: I0220 08:59:49.676754 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 08:59:49 crc kubenswrapper[4948]: W0220 08:59:49.711994 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1eaa8b0_3f84_4312_b7ac_4cc536093646.slice/crio-e5b803c8639b19bba9024b6977667b46b47a8a4045240c7e8378907b0fb57c9b WatchSource:0}: Error finding container e5b803c8639b19bba9024b6977667b46b47a8a4045240c7e8378907b0fb57c9b: Status 404 returned error can't find the container with id e5b803c8639b19bba9024b6977667b46b47a8a4045240c7e8378907b0fb57c9b Feb 20 08:59:50 crc kubenswrapper[4948]: I0220 08:59:50.236075 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" event={"ID":"e1eaa8b0-3f84-4312-b7ac-4cc536093646","Type":"ContainerStarted","Data":"e5b803c8639b19bba9024b6977667b46b47a8a4045240c7e8378907b0fb57c9b"} Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.143942 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn"] Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.146107 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.148665 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.149065 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.166451 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn"] Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.220283 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce5ff298-c730-4adc-ada7-e7a7268375e7-config-volume\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.220435 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce5ff298-c730-4adc-ada7-e7a7268375e7-secret-volume\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.220610 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwqlj\" (UniqueName: \"kubernetes.io/projected/ce5ff298-c730-4adc-ada7-e7a7268375e7-kube-api-access-mwqlj\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.322523 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce5ff298-c730-4adc-ada7-e7a7268375e7-secret-volume\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.322641 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwqlj\" (UniqueName: \"kubernetes.io/projected/ce5ff298-c730-4adc-ada7-e7a7268375e7-kube-api-access-mwqlj\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.322744 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce5ff298-c730-4adc-ada7-e7a7268375e7-config-volume\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.323736 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce5ff298-c730-4adc-ada7-e7a7268375e7-config-volume\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.328630 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce5ff298-c730-4adc-ada7-e7a7268375e7-secret-volume\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.341520 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwqlj\" (UniqueName: \"kubernetes.io/projected/ce5ff298-c730-4adc-ada7-e7a7268375e7-kube-api-access-mwqlj\") pod \"collect-profiles-29526300-g6rdn\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.469174 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:00 crc kubenswrapper[4948]: I0220 09:00:00.944190 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn"] Feb 20 09:00:01 crc kubenswrapper[4948]: I0220 09:00:01.339185 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" event={"ID":"e1eaa8b0-3f84-4312-b7ac-4cc536093646","Type":"ContainerStarted","Data":"d2d5efbac326ed9a4b4067346df3dcafa9289ae45429a1d4be737ea540693745"} Feb 20 09:00:01 crc kubenswrapper[4948]: I0220 09:00:01.342376 4948 generic.go:334] "Generic (PLEG): container finished" podID="ce5ff298-c730-4adc-ada7-e7a7268375e7" containerID="4e2abb2ff04ec67cba02074718775169ca5ab5ec96463e007a4d91e3fbf0ccfd" exitCode=0 Feb 20 09:00:01 crc kubenswrapper[4948]: I0220 09:00:01.342423 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" event={"ID":"ce5ff298-c730-4adc-ada7-e7a7268375e7","Type":"ContainerDied","Data":"4e2abb2ff04ec67cba02074718775169ca5ab5ec96463e007a4d91e3fbf0ccfd"} Feb 20 09:00:01 crc kubenswrapper[4948]: I0220 09:00:01.342456 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" event={"ID":"ce5ff298-c730-4adc-ada7-e7a7268375e7","Type":"ContainerStarted","Data":"b6a7a22636d78659185764a6d7a65810da15f8f6fec186645cce11bc9540f7fe"} Feb 20 09:00:01 crc kubenswrapper[4948]: I0220 09:00:01.362116 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" podStartSLOduration=1.607918326 podStartE2EDuration="12.362089958s" podCreationTimestamp="2026-02-20 08:59:49 +0000 UTC" firstStartedPulling="2026-02-20 08:59:49.715442807 +0000 UTC m=+3238.689937627" lastFinishedPulling="2026-02-20 09:00:00.469614439 +0000 UTC m=+3249.444109259" observedRunningTime="2026-02-20 09:00:01.355996587 +0000 UTC m=+3250.330491417" watchObservedRunningTime="2026-02-20 09:00:01.362089958 +0000 UTC m=+3250.336584798" Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.761261 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.866083 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce5ff298-c730-4adc-ada7-e7a7268375e7-config-volume\") pod \"ce5ff298-c730-4adc-ada7-e7a7268375e7\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.866152 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwqlj\" (UniqueName: \"kubernetes.io/projected/ce5ff298-c730-4adc-ada7-e7a7268375e7-kube-api-access-mwqlj\") pod \"ce5ff298-c730-4adc-ada7-e7a7268375e7\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.866467 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce5ff298-c730-4adc-ada7-e7a7268375e7-secret-volume\") pod \"ce5ff298-c730-4adc-ada7-e7a7268375e7\" (UID: \"ce5ff298-c730-4adc-ada7-e7a7268375e7\") " Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.866869 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce5ff298-c730-4adc-ada7-e7a7268375e7-config-volume" (OuterVolumeSpecName: "config-volume") pod "ce5ff298-c730-4adc-ada7-e7a7268375e7" (UID: "ce5ff298-c730-4adc-ada7-e7a7268375e7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.868432 4948 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce5ff298-c730-4adc-ada7-e7a7268375e7-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.874417 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce5ff298-c730-4adc-ada7-e7a7268375e7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ce5ff298-c730-4adc-ada7-e7a7268375e7" (UID: "ce5ff298-c730-4adc-ada7-e7a7268375e7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.899150 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce5ff298-c730-4adc-ada7-e7a7268375e7-kube-api-access-mwqlj" (OuterVolumeSpecName: "kube-api-access-mwqlj") pod "ce5ff298-c730-4adc-ada7-e7a7268375e7" (UID: "ce5ff298-c730-4adc-ada7-e7a7268375e7"). InnerVolumeSpecName "kube-api-access-mwqlj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.970354 4948 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce5ff298-c730-4adc-ada7-e7a7268375e7-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:02 crc kubenswrapper[4948]: I0220 09:00:02.970392 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwqlj\" (UniqueName: \"kubernetes.io/projected/ce5ff298-c730-4adc-ada7-e7a7268375e7-kube-api-access-mwqlj\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:03 crc kubenswrapper[4948]: I0220 09:00:03.360126 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" event={"ID":"ce5ff298-c730-4adc-ada7-e7a7268375e7","Type":"ContainerDied","Data":"b6a7a22636d78659185764a6d7a65810da15f8f6fec186645cce11bc9540f7fe"} Feb 20 09:00:03 crc kubenswrapper[4948]: I0220 09:00:03.360175 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6a7a22636d78659185764a6d7a65810da15f8f6fec186645cce11bc9540f7fe" Feb 20 09:00:03 crc kubenswrapper[4948]: I0220 09:00:03.360148 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-g6rdn" Feb 20 09:00:03 crc kubenswrapper[4948]: I0220 09:00:03.842207 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9"] Feb 20 09:00:03 crc kubenswrapper[4948]: I0220 09:00:03.849963 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526255-62px9"] Feb 20 09:00:05 crc kubenswrapper[4948]: I0220 09:00:05.733382 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9db735d8-ae70-471b-9d86-3be100901c72" path="/var/lib/kubelet/pods/9db735d8-ae70-471b-9d86-3be100901c72/volumes" Feb 20 09:00:37 crc kubenswrapper[4948]: I0220 09:00:37.663226 4948 generic.go:334] "Generic (PLEG): container finished" podID="e1eaa8b0-3f84-4312-b7ac-4cc536093646" containerID="d2d5efbac326ed9a4b4067346df3dcafa9289ae45429a1d4be737ea540693745" exitCode=0 Feb 20 09:00:37 crc kubenswrapper[4948]: I0220 09:00:37.663337 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" event={"ID":"e1eaa8b0-3f84-4312-b7ac-4cc536093646","Type":"ContainerDied","Data":"d2d5efbac326ed9a4b4067346df3dcafa9289ae45429a1d4be737ea540693745"} Feb 20 09:00:38 crc kubenswrapper[4948]: I0220 09:00:38.776126 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 09:00:38 crc kubenswrapper[4948]: I0220 09:00:38.834668 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-2rr8w"] Feb 20 09:00:38 crc kubenswrapper[4948]: I0220 09:00:38.848280 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-2rr8w"] Feb 20 09:00:38 crc kubenswrapper[4948]: I0220 09:00:38.924072 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bp4zr\" (UniqueName: \"kubernetes.io/projected/e1eaa8b0-3f84-4312-b7ac-4cc536093646-kube-api-access-bp4zr\") pod \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\" (UID: \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\") " Feb 20 09:00:38 crc kubenswrapper[4948]: I0220 09:00:38.924123 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e1eaa8b0-3f84-4312-b7ac-4cc536093646-host\") pod \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\" (UID: \"e1eaa8b0-3f84-4312-b7ac-4cc536093646\") " Feb 20 09:00:38 crc kubenswrapper[4948]: I0220 09:00:38.924545 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1eaa8b0-3f84-4312-b7ac-4cc536093646-host" (OuterVolumeSpecName: "host") pod "e1eaa8b0-3f84-4312-b7ac-4cc536093646" (UID: "e1eaa8b0-3f84-4312-b7ac-4cc536093646"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 09:00:38 crc kubenswrapper[4948]: I0220 09:00:38.930071 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1eaa8b0-3f84-4312-b7ac-4cc536093646-kube-api-access-bp4zr" (OuterVolumeSpecName: "kube-api-access-bp4zr") pod "e1eaa8b0-3f84-4312-b7ac-4cc536093646" (UID: "e1eaa8b0-3f84-4312-b7ac-4cc536093646"). InnerVolumeSpecName "kube-api-access-bp4zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:00:39 crc kubenswrapper[4948]: I0220 09:00:39.026407 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bp4zr\" (UniqueName: \"kubernetes.io/projected/e1eaa8b0-3f84-4312-b7ac-4cc536093646-kube-api-access-bp4zr\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:39 crc kubenswrapper[4948]: I0220 09:00:39.026741 4948 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e1eaa8b0-3f84-4312-b7ac-4cc536093646-host\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:39 crc kubenswrapper[4948]: I0220 09:00:39.685750 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5b803c8639b19bba9024b6977667b46b47a8a4045240c7e8378907b0fb57c9b" Feb 20 09:00:39 crc kubenswrapper[4948]: I0220 09:00:39.685808 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-2rr8w" Feb 20 09:00:39 crc kubenswrapper[4948]: I0220 09:00:39.735598 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1eaa8b0-3f84-4312-b7ac-4cc536093646" path="/var/lib/kubelet/pods/e1eaa8b0-3f84-4312-b7ac-4cc536093646/volumes" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.020601 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-fnl8n"] Feb 20 09:00:40 crc kubenswrapper[4948]: E0220 09:00:40.021436 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1eaa8b0-3f84-4312-b7ac-4cc536093646" containerName="container-00" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.021453 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1eaa8b0-3f84-4312-b7ac-4cc536093646" containerName="container-00" Feb 20 09:00:40 crc kubenswrapper[4948]: E0220 09:00:40.021477 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5ff298-c730-4adc-ada7-e7a7268375e7" containerName="collect-profiles" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.021485 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5ff298-c730-4adc-ada7-e7a7268375e7" containerName="collect-profiles" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.021715 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce5ff298-c730-4adc-ada7-e7a7268375e7" containerName="collect-profiles" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.021735 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1eaa8b0-3f84-4312-b7ac-4cc536093646" containerName="container-00" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.022450 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.024921 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-8sqsd"/"default-dockercfg-mzzkw" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.147605 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ngvx\" (UniqueName: \"kubernetes.io/projected/2116283d-53f0-45f3-9d0a-43f1af3d31d3-kube-api-access-8ngvx\") pod \"crc-debug-fnl8n\" (UID: \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\") " pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.147666 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2116283d-53f0-45f3-9d0a-43f1af3d31d3-host\") pod \"crc-debug-fnl8n\" (UID: \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\") " pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.249648 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ngvx\" (UniqueName: \"kubernetes.io/projected/2116283d-53f0-45f3-9d0a-43f1af3d31d3-kube-api-access-8ngvx\") pod \"crc-debug-fnl8n\" (UID: \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\") " pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.249702 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2116283d-53f0-45f3-9d0a-43f1af3d31d3-host\") pod \"crc-debug-fnl8n\" (UID: \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\") " pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.249855 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2116283d-53f0-45f3-9d0a-43f1af3d31d3-host\") pod \"crc-debug-fnl8n\" (UID: \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\") " pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.274358 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ngvx\" (UniqueName: \"kubernetes.io/projected/2116283d-53f0-45f3-9d0a-43f1af3d31d3-kube-api-access-8ngvx\") pod \"crc-debug-fnl8n\" (UID: \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\") " pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.337577 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.695855 4948 generic.go:334] "Generic (PLEG): container finished" podID="2116283d-53f0-45f3-9d0a-43f1af3d31d3" containerID="c7e69ff578f0981df401ec9cf440f5ada90becff9fd036ae8616d13d44ef4205" exitCode=0 Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.695931 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" event={"ID":"2116283d-53f0-45f3-9d0a-43f1af3d31d3","Type":"ContainerDied","Data":"c7e69ff578f0981df401ec9cf440f5ada90becff9fd036ae8616d13d44ef4205"} Feb 20 09:00:40 crc kubenswrapper[4948]: I0220 09:00:40.696233 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" event={"ID":"2116283d-53f0-45f3-9d0a-43f1af3d31d3","Type":"ContainerStarted","Data":"81333fd279dddbfd5f99a192bc9a5787fce10efd0531cd1b913b8e7cb477ce51"} Feb 20 09:00:41 crc kubenswrapper[4948]: I0220 09:00:41.108301 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-fnl8n"] Feb 20 09:00:41 crc kubenswrapper[4948]: I0220 09:00:41.118115 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-fnl8n"] Feb 20 09:00:41 crc kubenswrapper[4948]: I0220 09:00:41.815415 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:41 crc kubenswrapper[4948]: I0220 09:00:41.977848 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2116283d-53f0-45f3-9d0a-43f1af3d31d3-host\") pod \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\" (UID: \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\") " Feb 20 09:00:41 crc kubenswrapper[4948]: I0220 09:00:41.978043 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ngvx\" (UniqueName: \"kubernetes.io/projected/2116283d-53f0-45f3-9d0a-43f1af3d31d3-kube-api-access-8ngvx\") pod \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\" (UID: \"2116283d-53f0-45f3-9d0a-43f1af3d31d3\") " Feb 20 09:00:41 crc kubenswrapper[4948]: I0220 09:00:41.977955 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2116283d-53f0-45f3-9d0a-43f1af3d31d3-host" (OuterVolumeSpecName: "host") pod "2116283d-53f0-45f3-9d0a-43f1af3d31d3" (UID: "2116283d-53f0-45f3-9d0a-43f1af3d31d3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 09:00:41 crc kubenswrapper[4948]: I0220 09:00:41.978504 4948 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2116283d-53f0-45f3-9d0a-43f1af3d31d3-host\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:41 crc kubenswrapper[4948]: I0220 09:00:41.982807 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2116283d-53f0-45f3-9d0a-43f1af3d31d3-kube-api-access-8ngvx" (OuterVolumeSpecName: "kube-api-access-8ngvx") pod "2116283d-53f0-45f3-9d0a-43f1af3d31d3" (UID: "2116283d-53f0-45f3-9d0a-43f1af3d31d3"). InnerVolumeSpecName "kube-api-access-8ngvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.080434 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ngvx\" (UniqueName: \"kubernetes.io/projected/2116283d-53f0-45f3-9d0a-43f1af3d31d3-kube-api-access-8ngvx\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.291855 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-dbvll"] Feb 20 09:00:42 crc kubenswrapper[4948]: E0220 09:00:42.293352 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2116283d-53f0-45f3-9d0a-43f1af3d31d3" containerName="container-00" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.293484 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="2116283d-53f0-45f3-9d0a-43f1af3d31d3" containerName="container-00" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.293819 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="2116283d-53f0-45f3-9d0a-43f1af3d31d3" containerName="container-00" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.294651 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.387036 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f9bq\" (UniqueName: \"kubernetes.io/projected/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-kube-api-access-7f9bq\") pod \"crc-debug-dbvll\" (UID: \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\") " pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.387354 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-host\") pod \"crc-debug-dbvll\" (UID: \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\") " pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.489555 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f9bq\" (UniqueName: \"kubernetes.io/projected/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-kube-api-access-7f9bq\") pod \"crc-debug-dbvll\" (UID: \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\") " pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.489647 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-host\") pod \"crc-debug-dbvll\" (UID: \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\") " pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.489831 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-host\") pod \"crc-debug-dbvll\" (UID: \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\") " pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.505248 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f9bq\" (UniqueName: \"kubernetes.io/projected/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-kube-api-access-7f9bq\") pod \"crc-debug-dbvll\" (UID: \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\") " pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.609661 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:42 crc kubenswrapper[4948]: W0220 09:00:42.631820 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9fba2ec_0e22_4b23_aaa6_0406521ff59e.slice/crio-ffaf11f2cf3db38a687d2ffbc7c361de400b4facf9ae2cf5417b1b215c46e524 WatchSource:0}: Error finding container ffaf11f2cf3db38a687d2ffbc7c361de400b4facf9ae2cf5417b1b215c46e524: Status 404 returned error can't find the container with id ffaf11f2cf3db38a687d2ffbc7c361de400b4facf9ae2cf5417b1b215c46e524 Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.713106 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/crc-debug-dbvll" event={"ID":"e9fba2ec-0e22-4b23-aaa6-0406521ff59e","Type":"ContainerStarted","Data":"ffaf11f2cf3db38a687d2ffbc7c361de400b4facf9ae2cf5417b1b215c46e524"} Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.715035 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-fnl8n" Feb 20 09:00:42 crc kubenswrapper[4948]: I0220 09:00:42.715031 4948 scope.go:117] "RemoveContainer" containerID="c7e69ff578f0981df401ec9cf440f5ada90becff9fd036ae8616d13d44ef4205" Feb 20 09:00:43 crc kubenswrapper[4948]: I0220 09:00:43.725409 4948 generic.go:334] "Generic (PLEG): container finished" podID="e9fba2ec-0e22-4b23-aaa6-0406521ff59e" containerID="a21d252ab88ad9dc35d3fb9b8bd4786641c231e5c515a0b9849eb5e01174f261" exitCode=0 Feb 20 09:00:43 crc kubenswrapper[4948]: I0220 09:00:43.738927 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2116283d-53f0-45f3-9d0a-43f1af3d31d3" path="/var/lib/kubelet/pods/2116283d-53f0-45f3-9d0a-43f1af3d31d3/volumes" Feb 20 09:00:43 crc kubenswrapper[4948]: I0220 09:00:43.739774 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/crc-debug-dbvll" event={"ID":"e9fba2ec-0e22-4b23-aaa6-0406521ff59e","Type":"ContainerDied","Data":"a21d252ab88ad9dc35d3fb9b8bd4786641c231e5c515a0b9849eb5e01174f261"} Feb 20 09:00:43 crc kubenswrapper[4948]: I0220 09:00:43.775814 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-dbvll"] Feb 20 09:00:43 crc kubenswrapper[4948]: I0220 09:00:43.788650 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8sqsd/crc-debug-dbvll"] Feb 20 09:00:44 crc kubenswrapper[4948]: I0220 09:00:44.854095 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:44 crc kubenswrapper[4948]: I0220 09:00:44.940875 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-host\") pod \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\" (UID: \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\") " Feb 20 09:00:44 crc kubenswrapper[4948]: I0220 09:00:44.940964 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7f9bq\" (UniqueName: \"kubernetes.io/projected/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-kube-api-access-7f9bq\") pod \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\" (UID: \"e9fba2ec-0e22-4b23-aaa6-0406521ff59e\") " Feb 20 09:00:44 crc kubenswrapper[4948]: I0220 09:00:44.941030 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-host" (OuterVolumeSpecName: "host") pod "e9fba2ec-0e22-4b23-aaa6-0406521ff59e" (UID: "e9fba2ec-0e22-4b23-aaa6-0406521ff59e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 09:00:44 crc kubenswrapper[4948]: I0220 09:00:44.941579 4948 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-host\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:44 crc kubenswrapper[4948]: I0220 09:00:44.955224 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-kube-api-access-7f9bq" (OuterVolumeSpecName: "kube-api-access-7f9bq") pod "e9fba2ec-0e22-4b23-aaa6-0406521ff59e" (UID: "e9fba2ec-0e22-4b23-aaa6-0406521ff59e"). InnerVolumeSpecName "kube-api-access-7f9bq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:00:45 crc kubenswrapper[4948]: I0220 09:00:45.043282 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7f9bq\" (UniqueName: \"kubernetes.io/projected/e9fba2ec-0e22-4b23-aaa6-0406521ff59e-kube-api-access-7f9bq\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:45 crc kubenswrapper[4948]: I0220 09:00:45.732786 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9fba2ec-0e22-4b23-aaa6-0406521ff59e" path="/var/lib/kubelet/pods/e9fba2ec-0e22-4b23-aaa6-0406521ff59e/volumes" Feb 20 09:00:45 crc kubenswrapper[4948]: I0220 09:00:45.747583 4948 scope.go:117] "RemoveContainer" containerID="a21d252ab88ad9dc35d3fb9b8bd4786641c231e5c515a0b9849eb5e01174f261" Feb 20 09:00:45 crc kubenswrapper[4948]: I0220 09:00:45.747608 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/crc-debug-dbvll" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.172307 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-686d666564-8tnjb_51f6301c-57be-4a07-91df-b1a9f90e3f40/barbican-api/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.325096 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-686d666564-8tnjb_51f6301c-57be-4a07-91df-b1a9f90e3f40/barbican-api-log/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.398017 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6ffd468-xgk27_292d8e1e-de3b-4259-afef-c19e2c21187c/barbican-keystone-listener/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.438356 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6ffd468-xgk27_292d8e1e-de3b-4259-afef-c19e2c21187c/barbican-keystone-listener-log/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.571773 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76d486c65f-w9lc6_4b356058-d957-4fa5-a106-3fe5052c8539/barbican-worker/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.585502 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76d486c65f-w9lc6_4b356058-d957-4fa5-a106-3fe5052c8539/barbican-worker-log/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.742327 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-4265g_e9bed94f-dc85-433b-b2db-8da400959f54/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.792003 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58f563bb-9716-4abd-b187-3f01036d4f31/ceilometer-central-agent/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.883258 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58f563bb-9716-4abd-b187-3f01036d4f31/ceilometer-notification-agent/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.906626 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58f563bb-9716-4abd-b187-3f01036d4f31/proxy-httpd/0.log" Feb 20 09:00:59 crc kubenswrapper[4948]: I0220 09:00:59.936912 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58f563bb-9716-4abd-b187-3f01036d4f31/sg-core/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.069874 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e5967cab-cc29-4c87-b873-0f5d20203d4f/cinder-api/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.105335 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e5967cab-cc29-4c87-b873-0f5d20203d4f/cinder-api-log/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.146472 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29526301-gklw9"] Feb 20 09:01:00 crc kubenswrapper[4948]: E0220 09:01:00.147088 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9fba2ec-0e22-4b23-aaa6-0406521ff59e" containerName="container-00" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.147121 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9fba2ec-0e22-4b23-aaa6-0406521ff59e" containerName="container-00" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.147423 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9fba2ec-0e22-4b23-aaa6-0406521ff59e" containerName="container-00" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.148218 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.186273 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29526301-gklw9"] Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.259634 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-combined-ca-bundle\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.259766 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxrlv\" (UniqueName: \"kubernetes.io/projected/7990f62f-cd0f-4be2-b166-c00621d877e1-kube-api-access-pxrlv\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.259864 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-fernet-keys\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.259913 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-config-data\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.344331 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2f8a0f7b-ee05-4e93-a591-7486fecf8cd0/probe/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.346092 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2f8a0f7b-ee05-4e93-a591-7486fecf8cd0/cinder-scheduler/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.361046 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxrlv\" (UniqueName: \"kubernetes.io/projected/7990f62f-cd0f-4be2-b166-c00621d877e1-kube-api-access-pxrlv\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.361115 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-fernet-keys\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.361150 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-config-data\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.361225 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-combined-ca-bundle\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.367023 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-combined-ca-bundle\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.369401 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-config-data\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.381502 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-fernet-keys\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.381875 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxrlv\" (UniqueName: \"kubernetes.io/projected/7990f62f-cd0f-4be2-b166-c00621d877e1-kube-api-access-pxrlv\") pod \"keystone-cron-29526301-gklw9\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.390939 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx_bf3501bf-8f1e-4529-91a8-6be83eda4158/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.513359 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.622314 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-fbjpk_b05f92c6-60c9-41be-bf6c-6288cee04659/init/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.642433 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6_d38195a4-0c7b-427e-b5e2-923b24c10674/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.818224 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-fbjpk_b05f92c6-60c9-41be-bf6c-6288cee04659/init/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.896363 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-mhkll_5864a68d-650e-4bcf-b705-619c0f27445b/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.899506 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-fbjpk_b05f92c6-60c9-41be-bf6c-6288cee04659/dnsmasq-dns/0.log" Feb 20 09:01:00 crc kubenswrapper[4948]: I0220 09:01:00.961305 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29526301-gklw9"] Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.086848 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a/glance-log/0.log" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.116530 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a/glance-httpd/0.log" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.264467 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4fcb965e-4d04-4863-a966-39a83f458fa6/glance-httpd/0.log" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.281574 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4fcb965e-4d04-4863-a966-39a83f458fa6/glance-log/0.log" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.427722 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-685c4db87b-rw57h_7b9a072a-a040-414c-96c7-cf1454148745/horizon/0.log" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.592489 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw_c517042a-5354-4a8d-b7ff-c9aafa263b6c/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.694513 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-685c4db87b-rw57h_7b9a072a-a040-414c-96c7-cf1454148745/horizon-log/0.log" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.765884 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-4l55v_f42ddd0c-a2ec-414a-812d-f87213a3226d/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.883543 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526301-gklw9" event={"ID":"7990f62f-cd0f-4be2-b166-c00621d877e1","Type":"ContainerStarted","Data":"62da2ba3fadb7d2ed292d6ef544e7a030f473562d41eff9af8ce7a20a18cc887"} Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.883593 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526301-gklw9" event={"ID":"7990f62f-cd0f-4be2-b166-c00621d877e1","Type":"ContainerStarted","Data":"354be5b8ab4377e115ef71435a00840f8779e3587ce24ad7e77f1b79ea13775a"} Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.905262 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29526301-gklw9" podStartSLOduration=1.905245212 podStartE2EDuration="1.905245212s" podCreationTimestamp="2026-02-20 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 09:01:01.900883974 +0000 UTC m=+3310.875378794" watchObservedRunningTime="2026-02-20 09:01:01.905245212 +0000 UTC m=+3310.879740032" Feb 20 09:01:01 crc kubenswrapper[4948]: I0220 09:01:01.996915 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_24bf0fb4-81db-4955-8b5f-92c0f6b0856b/kube-state-metrics/0.log" Feb 20 09:01:02 crc kubenswrapper[4948]: I0220 09:01:02.041128 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6d4b8df464-wspg2_6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc/keystone-api/0.log" Feb 20 09:01:02 crc kubenswrapper[4948]: I0220 09:01:02.290214 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs_6e619eac-dfdc-4c8c-aa51-fc08ba58b49a/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:02 crc kubenswrapper[4948]: I0220 09:01:02.660619 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-76b5684765-h5624_fb9d5a3e-b07f-4229-9b5b-d0db6750cf59/neutron-api/0.log" Feb 20 09:01:02 crc kubenswrapper[4948]: I0220 09:01:02.786752 4948 scope.go:117] "RemoveContainer" containerID="984bcec5b88d4682af52feec68d576f83622efc6d9578f02504c68786eff5cfd" Feb 20 09:01:02 crc kubenswrapper[4948]: I0220 09:01:02.803357 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-76b5684765-h5624_fb9d5a3e-b07f-4229-9b5b-d0db6750cf59/neutron-httpd/0.log" Feb 20 09:01:02 crc kubenswrapper[4948]: I0220 09:01:02.942448 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr_fd280342-4276-4eee-a763-876e31be28c5/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:03 crc kubenswrapper[4948]: I0220 09:01:03.504550 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_873a6553-9637-4f2a-a743-f33cde7fc883/nova-cell0-conductor-conductor/0.log" Feb 20 09:01:03 crc kubenswrapper[4948]: I0220 09:01:03.506508 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_354f9b2f-9eaa-4e47-b214-0fd895f90c4c/nova-api-log/0.log" Feb 20 09:01:03 crc kubenswrapper[4948]: I0220 09:01:03.664898 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_354f9b2f-9eaa-4e47-b214-0fd895f90c4c/nova-api-api/0.log" Feb 20 09:01:03 crc kubenswrapper[4948]: I0220 09:01:03.834059 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_e801ada9-0247-4b29-b262-04637e1f8452/nova-cell1-conductor-conductor/0.log" Feb 20 09:01:03 crc kubenswrapper[4948]: I0220 09:01:03.853702 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9c00f24b-fedb-4747-91c4-3d0b551a7288/nova-cell1-novncproxy-novncproxy/0.log" Feb 20 09:01:03 crc kubenswrapper[4948]: I0220 09:01:03.907027 4948 generic.go:334] "Generic (PLEG): container finished" podID="7990f62f-cd0f-4be2-b166-c00621d877e1" containerID="62da2ba3fadb7d2ed292d6ef544e7a030f473562d41eff9af8ce7a20a18cc887" exitCode=0 Feb 20 09:01:03 crc kubenswrapper[4948]: I0220 09:01:03.907067 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526301-gklw9" event={"ID":"7990f62f-cd0f-4be2-b166-c00621d877e1","Type":"ContainerDied","Data":"62da2ba3fadb7d2ed292d6ef544e7a030f473562d41eff9af8ce7a20a18cc887"} Feb 20 09:01:04 crc kubenswrapper[4948]: I0220 09:01:04.076170 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-p22xr_6005ded0-9e21-4908-bbf0-33c710ba4341/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:04 crc kubenswrapper[4948]: I0220 09:01:04.148891 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_59965363-9971-4130-bb5a-5fcf5a44e2d5/nova-metadata-log/0.log" Feb 20 09:01:04 crc kubenswrapper[4948]: I0220 09:01:04.470643 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_f114bc41-01a5-4955-97eb-7fcf139cc5a9/nova-scheduler-scheduler/0.log" Feb 20 09:01:04 crc kubenswrapper[4948]: I0220 09:01:04.486124 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d02114ae-21f2-4cc5-9bca-9505c25eaf5f/mysql-bootstrap/0.log" Feb 20 09:01:04 crc kubenswrapper[4948]: I0220 09:01:04.650660 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d02114ae-21f2-4cc5-9bca-9505c25eaf5f/mysql-bootstrap/0.log" Feb 20 09:01:04 crc kubenswrapper[4948]: I0220 09:01:04.670220 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d02114ae-21f2-4cc5-9bca-9505c25eaf5f/galera/0.log" Feb 20 09:01:04 crc kubenswrapper[4948]: I0220 09:01:04.868310 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e59b63d-0c56-488c-87cb-348af87058c6/mysql-bootstrap/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.141141 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e59b63d-0c56-488c-87cb-348af87058c6/mysql-bootstrap/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.148557 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e59b63d-0c56-488c-87cb-348af87058c6/galera/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.237881 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.244427 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_59965363-9971-4130-bb5a-5fcf5a44e2d5/nova-metadata-metadata/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.315043 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_b2a76f68-3b4b-4993-aa4f-3b31b4ee1008/openstackclient/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.357187 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxrlv\" (UniqueName: \"kubernetes.io/projected/7990f62f-cd0f-4be2-b166-c00621d877e1-kube-api-access-pxrlv\") pod \"7990f62f-cd0f-4be2-b166-c00621d877e1\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.357248 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-fernet-keys\") pod \"7990f62f-cd0f-4be2-b166-c00621d877e1\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.357331 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-combined-ca-bundle\") pod \"7990f62f-cd0f-4be2-b166-c00621d877e1\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.357365 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-config-data\") pod \"7990f62f-cd0f-4be2-b166-c00621d877e1\" (UID: \"7990f62f-cd0f-4be2-b166-c00621d877e1\") " Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.375389 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7990f62f-cd0f-4be2-b166-c00621d877e1-kube-api-access-pxrlv" (OuterVolumeSpecName: "kube-api-access-pxrlv") pod "7990f62f-cd0f-4be2-b166-c00621d877e1" (UID: "7990f62f-cd0f-4be2-b166-c00621d877e1"). InnerVolumeSpecName "kube-api-access-pxrlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.375489 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7990f62f-cd0f-4be2-b166-c00621d877e1" (UID: "7990f62f-cd0f-4be2-b166-c00621d877e1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.411073 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-config-data" (OuterVolumeSpecName: "config-data") pod "7990f62f-cd0f-4be2-b166-c00621d877e1" (UID: "7990f62f-cd0f-4be2-b166-c00621d877e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.446180 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-gcfn8_1510e8b2-af6f-4641-a2d2-361830382250/openstack-network-exporter/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.460786 4948 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.460811 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxrlv\" (UniqueName: \"kubernetes.io/projected/7990f62f-cd0f-4be2-b166-c00621d877e1-kube-api-access-pxrlv\") on node \"crc\" DevicePath \"\"" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.460821 4948 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.478298 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7990f62f-cd0f-4be2-b166-c00621d877e1" (UID: "7990f62f-cd0f-4be2-b166-c00621d877e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.563092 4948 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7990f62f-cd0f-4be2-b166-c00621d877e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.756277 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q4bv4_c82772b6-9737-45ee-998a-e897086c03b5/ovsdb-server-init/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.766266 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-mkpng_b80cb988-de2a-4e65-9161-a0af0561c754/ovn-controller/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.922855 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q4bv4_c82772b6-9737-45ee-998a-e897086c03b5/ovsdb-server-init/0.log" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.926376 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526301-gklw9" event={"ID":"7990f62f-cd0f-4be2-b166-c00621d877e1","Type":"ContainerDied","Data":"354be5b8ab4377e115ef71435a00840f8779e3587ce24ad7e77f1b79ea13775a"} Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.926414 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="354be5b8ab4377e115ef71435a00840f8779e3587ce24ad7e77f1b79ea13775a" Feb 20 09:01:05 crc kubenswrapper[4948]: I0220 09:01:05.926467 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526301-gklw9" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.025946 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q4bv4_c82772b6-9737-45ee-998a-e897086c03b5/ovs-vswitchd/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.072718 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q4bv4_c82772b6-9737-45ee-998a-e897086c03b5/ovsdb-server/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.152071 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-4b2vp_b19ca7cb-01d2-4965-90f6-5d50211870cc/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.234052 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_eeb3782c-0a88-4d17-b481-e2aef7ea386a/openstack-network-exporter/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.280569 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_eeb3782c-0a88-4d17-b481-e2aef7ea386a/ovn-northd/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.424119 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d681442e-6bdb-46e4-9b49-6466e5e036a0/openstack-network-exporter/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.449054 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d681442e-6bdb-46e4-9b49-6466e5e036a0/ovsdbserver-nb/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.579119 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4b04e19d-cd3d-474d-884d-df12d0e1fffd/openstack-network-exporter/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.715503 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4b04e19d-cd3d-474d-884d-df12d0e1fffd/ovsdbserver-sb/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.785647 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-86887846d-pqjcb_5047ff00-845b-459f-ac89-6cbf997f93fb/placement-api/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.863795 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-86887846d-pqjcb_5047ff00-845b-459f-ac89-6cbf997f93fb/placement-log/0.log" Feb 20 09:01:06 crc kubenswrapper[4948]: I0220 09:01:06.921845 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_037b329f-712f-4a67-984a-75affd2a57b7/setup-container/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.154306 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_037b329f-712f-4a67-984a-75affd2a57b7/setup-container/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.178654 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_037b329f-712f-4a67-984a-75affd2a57b7/rabbitmq/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.319041 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f063bb62-10c3-46d9-a41d-56d2826059e0/setup-container/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.475576 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f063bb62-10c3-46d9-a41d-56d2826059e0/setup-container/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.520684 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f063bb62-10c3-46d9-a41d-56d2826059e0/rabbitmq/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.561577 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-ds522_e3b05342-f970-4fdb-9b1c-ff521e118f96/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.758141 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-f7ww9_c3e7cce4-5144-4190-8013-83428c66cde9/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.787745 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch_90b6e7df-b3e1-4d05-a6b3-7fe47d70804f/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:07 crc kubenswrapper[4948]: I0220 09:01:07.973877 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-lkbkf_79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.024316 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.024369 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.042922 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-cqchc_41856163-b621-439a-95c1-ca8d13f0a08c/ssh-known-hosts-edpm-deployment/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.277957 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7c7cd9876c-262qf_30f48966-e44d-44bc-a868-7e8119ac186c/proxy-server/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.281334 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7c7cd9876c-262qf_30f48966-e44d-44bc-a868-7e8119ac186c/proxy-httpd/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.410903 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-dllls_642adfd1-e253-4e0a-80e6-860a9f5c1ae1/swift-ring-rebalance/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.516230 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/account-auditor/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.596597 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/account-reaper/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.679164 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/account-replicator/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.698267 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/account-server/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.755753 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/container-auditor/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.849790 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/container-replicator/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.875222 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/container-server/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.914707 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/container-updater/0.log" Feb 20 09:01:08 crc kubenswrapper[4948]: I0220 09:01:08.974195 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-auditor/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.044261 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-expirer/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.079439 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-replicator/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.107811 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-server/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.166741 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-updater/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.265192 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/rsync/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.344093 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/swift-recon-cron/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.478441 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd_06ed5839-bc31-4691-8586-cd42c0413006/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.580106 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_0fefa0c4-ece9-4caf-8b41-8a64472c53e6/tempest-tests-tempest-tests-runner/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.699114 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_de083766-efe0-42cd-95c5-b87bbeb33b0a/test-operator-logs-container/0.log" Feb 20 09:01:09 crc kubenswrapper[4948]: I0220 09:01:09.825959 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-lf98d_2e05be38-f4ab-415f-a71d-9e5233cf0ea7/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:01:13 crc kubenswrapper[4948]: I0220 09:01:13.967312 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-7c7cd9876c-262qf" podUID="30f48966-e44d-44bc-a868-7e8119ac186c" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Feb 20 09:01:19 crc kubenswrapper[4948]: I0220 09:01:19.084404 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_8e994489-a2cf-4f6d-a00c-98f627ba0e5f/memcached/0.log" Feb 20 09:01:35 crc kubenswrapper[4948]: I0220 09:01:35.013468 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/util/0.log" Feb 20 09:01:35 crc kubenswrapper[4948]: I0220 09:01:35.212052 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/pull/0.log" Feb 20 09:01:35 crc kubenswrapper[4948]: I0220 09:01:35.268492 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/pull/0.log" Feb 20 09:01:35 crc kubenswrapper[4948]: I0220 09:01:35.271947 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/util/0.log" Feb 20 09:01:35 crc kubenswrapper[4948]: I0220 09:01:35.472118 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/pull/0.log" Feb 20 09:01:35 crc kubenswrapper[4948]: I0220 09:01:35.489067 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/extract/0.log" Feb 20 09:01:35 crc kubenswrapper[4948]: I0220 09:01:35.494681 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/util/0.log" Feb 20 09:01:35 crc kubenswrapper[4948]: I0220 09:01:35.872628 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-kjpkr_98a1aa06-948b-4034-bc07-7e546e341a8f/manager/0.log" Feb 20 09:01:36 crc kubenswrapper[4948]: I0220 09:01:36.234317 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987464f4-xc74s_347cdd7c-e5e8-49b1-a4f1-687b8a06b250/manager/0.log" Feb 20 09:01:36 crc kubenswrapper[4948]: I0220 09:01:36.370850 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-6cr5p_2d303d1a-1062-4bf6-be68-1bd6d1a3228f/manager/0.log" Feb 20 09:01:36 crc kubenswrapper[4948]: I0220 09:01:36.585072 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-bkl49_4a663826-d4b5-4ed3-8270-099b003390b4/manager/0.log" Feb 20 09:01:36 crc kubenswrapper[4948]: I0220 09:01:36.939601 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5d946d989d-c6xqd_586113f2-38b6-4bd1-8adb-3c155bb35ba9/manager/0.log" Feb 20 09:01:37 crc kubenswrapper[4948]: I0220 09:01:37.080072 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-5fqzw_105a118c-121d-4582-960f-1da9957980cb/manager/0.log" Feb 20 09:01:37 crc kubenswrapper[4948]: I0220 09:01:37.120131 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79d975b745-jrlw5_93b93501-acfe-4274-9a22-ca644b1d11d3/manager/0.log" Feb 20 09:01:37 crc kubenswrapper[4948]: I0220 09:01:37.344521 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-54f6768c69-lj2sz_f762ec0d-e09c-4baf-9540-f4ac61ce7234/manager/0.log" Feb 20 09:01:37 crc kubenswrapper[4948]: I0220 09:01:37.366364 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-ttkfm_3fbbd11d-612d-479a-a34f-505d995a4871/manager/0.log" Feb 20 09:01:37 crc kubenswrapper[4948]: I0220 09:01:37.613620 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-zcccd_6e8e8b87-5a00-486f-a00c-1450c109f3b2/manager/0.log" Feb 20 09:01:37 crc kubenswrapper[4948]: I0220 09:01:37.875163 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64ddbf8bb-jcnxd_88142137-864d-4660-a688-a7dcc503851b/manager/0.log" Feb 20 09:01:37 crc kubenswrapper[4948]: I0220 09:01:37.927764 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-jxwsx_7c7bb531-7900-4cc6-9d9b-bae52dabc59d/manager/0.log" Feb 20 09:01:38 crc kubenswrapper[4948]: I0220 09:01:38.024302 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:01:38 crc kubenswrapper[4948]: I0220 09:01:38.024350 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:01:38 crc kubenswrapper[4948]: I0220 09:01:38.302866 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m_b8b07685-095c-4bbd-a30d-57ca59d7cbdc/manager/0.log" Feb 20 09:01:38 crc kubenswrapper[4948]: I0220 09:01:38.675483 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-59c446d8d6-4cs2n_9465a7a8-c08e-44c9-a76d-3878b0bff3af/operator/0.log" Feb 20 09:01:39 crc kubenswrapper[4948]: I0220 09:01:39.054817 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-9z4jj_64872326-fb95-4bc1-a6c5-0b34242883f9/registry-server/0.log" Feb 20 09:01:39 crc kubenswrapper[4948]: I0220 09:01:39.368806 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-d44cf6b75-dkkq9_3ea675de-d1b0-4880-9652-eb066f6b0fb7/manager/0.log" Feb 20 09:01:39 crc kubenswrapper[4948]: I0220 09:01:39.549209 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f8888797-hl5w7_ee33bd02-a3ce-415d-9d2d-fefd383e9810/manager/0.log" Feb 20 09:01:39 crc kubenswrapper[4948]: I0220 09:01:39.592378 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-znrkz_16933747-642c-45ff-9f98-9321c633826a/manager/0.log" Feb 20 09:01:39 crc kubenswrapper[4948]: I0220 09:01:39.763480 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-59pcs_fe0aca33-1acf-463a-91a6-bbf35a38fd7c/operator/0.log" Feb 20 09:01:39 crc kubenswrapper[4948]: I0220 09:01:39.840132 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-5lfts_d65e0993-11a0-4e81-963f-eeb9dcb92536/manager/0.log" Feb 20 09:01:40 crc kubenswrapper[4948]: I0220 09:01:40.051835 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f45b4ff68-5rrd9_0f0cbb65-3c8e-41e6-8059-34e121de0821/manager/0.log" Feb 20 09:01:40 crc kubenswrapper[4948]: I0220 09:01:40.133756 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7866795846-f7cxn_0f98bc58-d0e3-405b-88fd-d8bd65f415a4/manager/0.log" Feb 20 09:01:40 crc kubenswrapper[4948]: I0220 09:01:40.287257 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5db88f68c-pndq7_f2400173-aa87-476c-8216-4f8c9cf9d474/manager/0.log" Feb 20 09:01:40 crc kubenswrapper[4948]: I0220 09:01:40.365532 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8569bc6fdb-lt4jk_e9f13bcd-ac5a-4cbd-952a-a5b9dceee562/manager/0.log" Feb 20 09:01:42 crc kubenswrapper[4948]: I0220 09:01:42.216770 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-4h9v5_b8e12d0c-9564-4bf2-ac61-b22d2fbdf855/manager/0.log" Feb 20 09:01:59 crc kubenswrapper[4948]: I0220 09:01:59.487175 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-bgm85_a149fe37-c748-4120-9116-1da4b680d880/control-plane-machine-set-operator/0.log" Feb 20 09:01:59 crc kubenswrapper[4948]: I0220 09:01:59.686599 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fthdq_5bc3d0cd-eacd-4b25-9acd-853e49db7b47/kube-rbac-proxy/0.log" Feb 20 09:01:59 crc kubenswrapper[4948]: I0220 09:01:59.724991 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fthdq_5bc3d0cd-eacd-4b25-9acd-853e49db7b47/machine-api-operator/0.log" Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.024927 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.027230 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.027371 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.028247 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b0e4da4a6a513a5c47c1f0dd87dcf0b56d3fede7e9008d839ac5623f61075ff8"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.028409 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://b0e4da4a6a513a5c47c1f0dd87dcf0b56d3fede7e9008d839ac5623f61075ff8" gracePeriod=600 Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.547624 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="b0e4da4a6a513a5c47c1f0dd87dcf0b56d3fede7e9008d839ac5623f61075ff8" exitCode=0 Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.547808 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"b0e4da4a6a513a5c47c1f0dd87dcf0b56d3fede7e9008d839ac5623f61075ff8"} Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.548069 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2"} Feb 20 09:02:08 crc kubenswrapper[4948]: I0220 09:02:08.548098 4948 scope.go:117] "RemoveContainer" containerID="cfe1e07886883274c9cb7951efaafb3d233396b16e657f9b5c4c23b1283205a5" Feb 20 09:02:13 crc kubenswrapper[4948]: I0220 09:02:13.471769 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-smtqw_1437d818-abf4-4602-8a28-e88a76e482a9/cert-manager-controller/0.log" Feb 20 09:02:13 crc kubenswrapper[4948]: I0220 09:02:13.679166 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-f4bjc_305a17f4-aef3-4036-8fce-3756ff5bbd2f/cert-manager-webhook/0.log" Feb 20 09:02:13 crc kubenswrapper[4948]: I0220 09:02:13.682806 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-9llfv_bb56399a-a840-49af-972f-4f32c91efed7/cert-manager-cainjector/0.log" Feb 20 09:02:26 crc kubenswrapper[4948]: I0220 09:02:26.888552 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5c78fc5d65-9mv9z_47461ed8-c51b-4014-bcf0-4d95a0278f85/nmstate-console-plugin/0.log" Feb 20 09:02:27 crc kubenswrapper[4948]: I0220 09:02:27.086357 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-dxdh6_3a869b3a-b757-49ba-9096-2e562d980aae/nmstate-handler/0.log" Feb 20 09:02:27 crc kubenswrapper[4948]: I0220 09:02:27.166435 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-tdst7_a7b34c06-d0c5-4f20-9521-eec500ea2510/nmstate-metrics/0.log" Feb 20 09:02:27 crc kubenswrapper[4948]: I0220 09:02:27.179330 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-tdst7_a7b34c06-d0c5-4f20-9521-eec500ea2510/kube-rbac-proxy/0.log" Feb 20 09:02:27 crc kubenswrapper[4948]: I0220 09:02:27.323278 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-694c9596b7-xqtvg_59277281-0112-45da-a64a-1b1d02b90473/nmstate-operator/0.log" Feb 20 09:02:27 crc kubenswrapper[4948]: I0220 09:02:27.357952 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-866bcb46dc-77sdv_da0e4284-f20d-44a8-b529-e388f15822d3/nmstate-webhook/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.056653 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-c6sqk_68ee0185-13d4-45c7-9a39-bb474e9314d0/kube-rbac-proxy/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.088984 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-c6sqk_68ee0185-13d4-45c7-9a39-bb474e9314d0/controller/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.216315 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-frr-files/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.431943 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-reloader/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.439882 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-reloader/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.451433 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-frr-files/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.472885 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-metrics/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.604499 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-metrics/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.612797 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-frr-files/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.636472 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-reloader/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.679892 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-metrics/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.795033 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-reloader/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.816872 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-frr-files/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.822458 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-metrics/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.859885 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/controller/0.log" Feb 20 09:02:56 crc kubenswrapper[4948]: I0220 09:02:56.986285 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/frr-metrics/0.log" Feb 20 09:02:57 crc kubenswrapper[4948]: I0220 09:02:57.113910 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/kube-rbac-proxy-frr/0.log" Feb 20 09:02:57 crc kubenswrapper[4948]: I0220 09:02:57.122798 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/kube-rbac-proxy/0.log" Feb 20 09:02:57 crc kubenswrapper[4948]: I0220 09:02:57.234668 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/reloader/0.log" Feb 20 09:02:57 crc kubenswrapper[4948]: I0220 09:02:57.375166 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-7zsxs_2506d4e2-b434-4fe4-970e-7cd14601677d/frr-k8s-webhook-server/0.log" Feb 20 09:02:57 crc kubenswrapper[4948]: I0220 09:02:57.532751 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-8478c6b8cc-v98zs_b4f86a2e-15f9-441c-953b-49f331d4122e/manager/0.log" Feb 20 09:02:57 crc kubenswrapper[4948]: I0220 09:02:57.633246 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5fcd7d546c-kgpxs_07022b13-7dc1-45d1-a626-91792bd3aa90/webhook-server/0.log" Feb 20 09:02:57 crc kubenswrapper[4948]: I0220 09:02:57.832126 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-h7xhj_2c199acd-2fd3-4b9f-b50c-a266191c4777/kube-rbac-proxy/0.log" Feb 20 09:02:58 crc kubenswrapper[4948]: I0220 09:02:58.260698 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-h7xhj_2c199acd-2fd3-4b9f-b50c-a266191c4777/speaker/0.log" Feb 20 09:02:58 crc kubenswrapper[4948]: I0220 09:02:58.307790 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/frr/0.log" Feb 20 09:03:11 crc kubenswrapper[4948]: I0220 09:03:11.971511 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/util/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.177563 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/util/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.363171 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/pull/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.398645 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/pull/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.583427 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/pull/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.606452 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/extract/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.608145 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/util/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.768800 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-utilities/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.937222 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-content/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.964863 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-utilities/0.log" Feb 20 09:03:12 crc kubenswrapper[4948]: I0220 09:03:12.996850 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-content/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.124701 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-utilities/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.148464 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-content/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.366946 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-utilities/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.496734 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-content/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.508333 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/registry-server/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.529591 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-utilities/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.597669 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-content/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.727565 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-utilities/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.768493 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-content/0.log" Feb 20 09:03:13 crc kubenswrapper[4948]: I0220 09:03:13.972889 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/util/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.145133 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/util/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.151209 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/pull/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.249882 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/registry-server/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.258394 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/pull/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.440170 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/pull/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.454049 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/util/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.455472 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/extract/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.647526 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-utilities/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.721924 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-tbqzs_1e3d2cd4-4311-4454-b257-20a3caf243b3/marketplace-operator/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.859745 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-content/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.902903 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-utilities/0.log" Feb 20 09:03:14 crc kubenswrapper[4948]: I0220 09:03:14.917430 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-content/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.074691 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-utilities/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.077897 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-content/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.264540 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/registry-server/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.311456 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-utilities/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.473484 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-content/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.498000 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-content/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.505267 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-utilities/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.646567 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-utilities/0.log" Feb 20 09:03:15 crc kubenswrapper[4948]: I0220 09:03:15.711747 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-content/0.log" Feb 20 09:03:16 crc kubenswrapper[4948]: I0220 09:03:16.027306 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/registry-server/0.log" Feb 20 09:03:37 crc kubenswrapper[4948]: E0220 09:03:37.505654 4948 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.113:43138->38.102.83.113:38489: write tcp 38.102.83.113:43138->38.102.83.113:38489: write: broken pipe Feb 20 09:04:08 crc kubenswrapper[4948]: I0220 09:04:08.024915 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:04:08 crc kubenswrapper[4948]: I0220 09:04:08.025561 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.337836 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cqvgj"] Feb 20 09:04:34 crc kubenswrapper[4948]: E0220 09:04:34.339106 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7990f62f-cd0f-4be2-b166-c00621d877e1" containerName="keystone-cron" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.339132 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="7990f62f-cd0f-4be2-b166-c00621d877e1" containerName="keystone-cron" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.339421 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="7990f62f-cd0f-4be2-b166-c00621d877e1" containerName="keystone-cron" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.341522 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.352287 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cqvgj"] Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.413804 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-utilities\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.413860 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lphb2\" (UniqueName: \"kubernetes.io/projected/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-kube-api-access-lphb2\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.413940 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-catalog-content\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.515743 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-utilities\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.515782 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lphb2\" (UniqueName: \"kubernetes.io/projected/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-kube-api-access-lphb2\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.515843 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-catalog-content\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.516350 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-utilities\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.516416 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-catalog-content\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.539156 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lphb2\" (UniqueName: \"kubernetes.io/projected/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-kube-api-access-lphb2\") pod \"redhat-operators-cqvgj\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:34 crc kubenswrapper[4948]: I0220 09:04:34.688777 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:35 crc kubenswrapper[4948]: I0220 09:04:35.166396 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cqvgj"] Feb 20 09:04:36 crc kubenswrapper[4948]: I0220 09:04:36.010062 4948 generic.go:334] "Generic (PLEG): container finished" podID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerID="d54544788f251e4b6d29e4c674bc8416aca134867b24b018c4f49416d726ab6e" exitCode=0 Feb 20 09:04:36 crc kubenswrapper[4948]: I0220 09:04:36.010297 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqvgj" event={"ID":"fa0586ff-df65-4d4d-8571-eb1d0bed5b62","Type":"ContainerDied","Data":"d54544788f251e4b6d29e4c674bc8416aca134867b24b018c4f49416d726ab6e"} Feb 20 09:04:36 crc kubenswrapper[4948]: I0220 09:04:36.010389 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqvgj" event={"ID":"fa0586ff-df65-4d4d-8571-eb1d0bed5b62","Type":"ContainerStarted","Data":"e0a78829b49373d7761eae4a5b21ce062c4d2de49690ef10c9018d2c5d7916c2"} Feb 20 09:04:36 crc kubenswrapper[4948]: I0220 09:04:36.018150 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 09:04:37 crc kubenswrapper[4948]: I0220 09:04:37.022128 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqvgj" event={"ID":"fa0586ff-df65-4d4d-8571-eb1d0bed5b62","Type":"ContainerStarted","Data":"49c59846e9d076d7fd5056bdfdce13873c0177f11109eb4581cf9a83aaa33bdc"} Feb 20 09:04:38 crc kubenswrapper[4948]: I0220 09:04:38.024496 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:04:38 crc kubenswrapper[4948]: I0220 09:04:38.024569 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:04:39 crc kubenswrapper[4948]: I0220 09:04:39.040776 4948 generic.go:334] "Generic (PLEG): container finished" podID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerID="49c59846e9d076d7fd5056bdfdce13873c0177f11109eb4581cf9a83aaa33bdc" exitCode=0 Feb 20 09:04:39 crc kubenswrapper[4948]: I0220 09:04:39.040852 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqvgj" event={"ID":"fa0586ff-df65-4d4d-8571-eb1d0bed5b62","Type":"ContainerDied","Data":"49c59846e9d076d7fd5056bdfdce13873c0177f11109eb4581cf9a83aaa33bdc"} Feb 20 09:04:40 crc kubenswrapper[4948]: I0220 09:04:40.060766 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqvgj" event={"ID":"fa0586ff-df65-4d4d-8571-eb1d0bed5b62","Type":"ContainerStarted","Data":"aba938c4a927bdefe6c39509a204ce9987b639f0088daf4bdfdf4018402bc5e8"} Feb 20 09:04:40 crc kubenswrapper[4948]: I0220 09:04:40.097828 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cqvgj" podStartSLOduration=2.593158372 podStartE2EDuration="6.097802944s" podCreationTimestamp="2026-02-20 09:04:34 +0000 UTC" firstStartedPulling="2026-02-20 09:04:36.017915291 +0000 UTC m=+3524.992410111" lastFinishedPulling="2026-02-20 09:04:39.522559853 +0000 UTC m=+3528.497054683" observedRunningTime="2026-02-20 09:04:40.089275413 +0000 UTC m=+3529.063770273" watchObservedRunningTime="2026-02-20 09:04:40.097802944 +0000 UTC m=+3529.072297794" Feb 20 09:04:44 crc kubenswrapper[4948]: I0220 09:04:44.689554 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:44 crc kubenswrapper[4948]: I0220 09:04:44.691838 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:45 crc kubenswrapper[4948]: I0220 09:04:45.772575 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cqvgj" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="registry-server" probeResult="failure" output=< Feb 20 09:04:45 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 09:04:45 crc kubenswrapper[4948]: > Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.102549 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2hl84"] Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.108255 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.124436 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2hl84"] Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.281057 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hv8x\" (UniqueName: \"kubernetes.io/projected/27f74a12-0843-4c61-a5c5-7f026224d86e-kube-api-access-6hv8x\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.281145 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-catalog-content\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.281164 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-utilities\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.383197 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-catalog-content\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.383243 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-utilities\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.383370 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hv8x\" (UniqueName: \"kubernetes.io/projected/27f74a12-0843-4c61-a5c5-7f026224d86e-kube-api-access-6hv8x\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.383671 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-catalog-content\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.384053 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-utilities\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.402166 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hv8x\" (UniqueName: \"kubernetes.io/projected/27f74a12-0843-4c61-a5c5-7f026224d86e-kube-api-access-6hv8x\") pod \"certified-operators-2hl84\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:53 crc kubenswrapper[4948]: I0220 09:04:53.473022 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:04:54 crc kubenswrapper[4948]: I0220 09:04:54.026826 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2hl84"] Feb 20 09:04:54 crc kubenswrapper[4948]: I0220 09:04:54.199499 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2hl84" event={"ID":"27f74a12-0843-4c61-a5c5-7f026224d86e","Type":"ContainerStarted","Data":"9d10453c1bde550a3b492982c9301c05c35c72c85021ea721f6d12fcfa8a9df5"} Feb 20 09:04:54 crc kubenswrapper[4948]: I0220 09:04:54.789935 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:54 crc kubenswrapper[4948]: I0220 09:04:54.854201 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:55 crc kubenswrapper[4948]: I0220 09:04:55.215257 4948 generic.go:334] "Generic (PLEG): container finished" podID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerID="274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76" exitCode=0 Feb 20 09:04:55 crc kubenswrapper[4948]: I0220 09:04:55.215334 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2hl84" event={"ID":"27f74a12-0843-4c61-a5c5-7f026224d86e","Type":"ContainerDied","Data":"274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76"} Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.084678 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cqvgj"] Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.085454 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cqvgj" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="registry-server" containerID="cri-o://aba938c4a927bdefe6c39509a204ce9987b639f0088daf4bdfdf4018402bc5e8" gracePeriod=2 Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.234362 4948 generic.go:334] "Generic (PLEG): container finished" podID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerID="b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb" exitCode=0 Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.234433 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2hl84" event={"ID":"27f74a12-0843-4c61-a5c5-7f026224d86e","Type":"ContainerDied","Data":"b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb"} Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.243071 4948 generic.go:334] "Generic (PLEG): container finished" podID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerID="aba938c4a927bdefe6c39509a204ce9987b639f0088daf4bdfdf4018402bc5e8" exitCode=0 Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.243116 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqvgj" event={"ID":"fa0586ff-df65-4d4d-8571-eb1d0bed5b62","Type":"ContainerDied","Data":"aba938c4a927bdefe6c39509a204ce9987b639f0088daf4bdfdf4018402bc5e8"} Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.583487 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.679400 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-utilities\") pod \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.679501 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lphb2\" (UniqueName: \"kubernetes.io/projected/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-kube-api-access-lphb2\") pod \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.679674 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-catalog-content\") pod \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\" (UID: \"fa0586ff-df65-4d4d-8571-eb1d0bed5b62\") " Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.690878 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-utilities" (OuterVolumeSpecName: "utilities") pod "fa0586ff-df65-4d4d-8571-eb1d0bed5b62" (UID: "fa0586ff-df65-4d4d-8571-eb1d0bed5b62"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.701632 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-kube-api-access-lphb2" (OuterVolumeSpecName: "kube-api-access-lphb2") pod "fa0586ff-df65-4d4d-8571-eb1d0bed5b62" (UID: "fa0586ff-df65-4d4d-8571-eb1d0bed5b62"). InnerVolumeSpecName "kube-api-access-lphb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.782274 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lphb2\" (UniqueName: \"kubernetes.io/projected/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-kube-api-access-lphb2\") on node \"crc\" DevicePath \"\"" Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.782313 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.837251 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa0586ff-df65-4d4d-8571-eb1d0bed5b62" (UID: "fa0586ff-df65-4d4d-8571-eb1d0bed5b62"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:04:57 crc kubenswrapper[4948]: I0220 09:04:57.884927 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0586ff-df65-4d4d-8571-eb1d0bed5b62-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.268217 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2hl84" event={"ID":"27f74a12-0843-4c61-a5c5-7f026224d86e","Type":"ContainerStarted","Data":"42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca"} Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.271711 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqvgj" event={"ID":"fa0586ff-df65-4d4d-8571-eb1d0bed5b62","Type":"ContainerDied","Data":"e0a78829b49373d7761eae4a5b21ce062c4d2de49690ef10c9018d2c5d7916c2"} Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.271764 4948 scope.go:117] "RemoveContainer" containerID="aba938c4a927bdefe6c39509a204ce9987b639f0088daf4bdfdf4018402bc5e8" Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.271769 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cqvgj" Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.301586 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2hl84" podStartSLOduration=2.870851341 podStartE2EDuration="5.301560437s" podCreationTimestamp="2026-02-20 09:04:53 +0000 UTC" firstStartedPulling="2026-02-20 09:04:55.221587873 +0000 UTC m=+3544.196082743" lastFinishedPulling="2026-02-20 09:04:57.652297019 +0000 UTC m=+3546.626791839" observedRunningTime="2026-02-20 09:04:58.301396043 +0000 UTC m=+3547.275890893" watchObservedRunningTime="2026-02-20 09:04:58.301560437 +0000 UTC m=+3547.276055287" Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.323603 4948 scope.go:117] "RemoveContainer" containerID="49c59846e9d076d7fd5056bdfdce13873c0177f11109eb4581cf9a83aaa33bdc" Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.335867 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cqvgj"] Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.355931 4948 scope.go:117] "RemoveContainer" containerID="d54544788f251e4b6d29e4c674bc8416aca134867b24b018c4f49416d726ab6e" Feb 20 09:04:58 crc kubenswrapper[4948]: I0220 09:04:58.358464 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cqvgj"] Feb 20 09:04:59 crc kubenswrapper[4948]: I0220 09:04:59.737953 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" path="/var/lib/kubelet/pods/fa0586ff-df65-4d4d-8571-eb1d0bed5b62/volumes" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.084830 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gs7f6"] Feb 20 09:05:02 crc kubenswrapper[4948]: E0220 09:05:02.085634 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="extract-content" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.085648 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="extract-content" Feb 20 09:05:02 crc kubenswrapper[4948]: E0220 09:05:02.085665 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="registry-server" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.085671 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="registry-server" Feb 20 09:05:02 crc kubenswrapper[4948]: E0220 09:05:02.085689 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="extract-utilities" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.085696 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="extract-utilities" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.085877 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa0586ff-df65-4d4d-8571-eb1d0bed5b62" containerName="registry-server" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.087467 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.113569 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gs7f6"] Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.283192 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-catalog-content\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.283248 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-utilities\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.283457 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmpwq\" (UniqueName: \"kubernetes.io/projected/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-kube-api-access-bmpwq\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.322823 4948 generic.go:334] "Generic (PLEG): container finished" podID="b25c36d5-5de9-4c98-acf4-495d492def68" containerID="4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a" exitCode=0 Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.322868 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8sqsd/must-gather-xzw72" event={"ID":"b25c36d5-5de9-4c98-acf4-495d492def68","Type":"ContainerDied","Data":"4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a"} Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.323505 4948 scope.go:117] "RemoveContainer" containerID="4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.385836 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-catalog-content\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.385876 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-utilities\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.385929 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmpwq\" (UniqueName: \"kubernetes.io/projected/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-kube-api-access-bmpwq\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.386517 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-catalog-content\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.386579 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-utilities\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.407958 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmpwq\" (UniqueName: \"kubernetes.io/projected/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-kube-api-access-bmpwq\") pod \"community-operators-gs7f6\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.428133 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:02 crc kubenswrapper[4948]: I0220 09:05:02.939727 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gs7f6"] Feb 20 09:05:02 crc kubenswrapper[4948]: W0220 09:05:02.950236 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb1bc246_7f5a_4060_8bdb_6a3963ea6862.slice/crio-a09d366d7a1bea20bacda46f1149340ebdb8bdbac35df3dd60174f0f350bbfa7 WatchSource:0}: Error finding container a09d366d7a1bea20bacda46f1149340ebdb8bdbac35df3dd60174f0f350bbfa7: Status 404 returned error can't find the container with id a09d366d7a1bea20bacda46f1149340ebdb8bdbac35df3dd60174f0f350bbfa7 Feb 20 09:05:03 crc kubenswrapper[4948]: I0220 09:05:03.102552 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8sqsd_must-gather-xzw72_b25c36d5-5de9-4c98-acf4-495d492def68/gather/0.log" Feb 20 09:05:03 crc kubenswrapper[4948]: I0220 09:05:03.335615 4948 generic.go:334] "Generic (PLEG): container finished" podID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerID="ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685" exitCode=0 Feb 20 09:05:03 crc kubenswrapper[4948]: I0220 09:05:03.335657 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs7f6" event={"ID":"fb1bc246-7f5a-4060-8bdb-6a3963ea6862","Type":"ContainerDied","Data":"ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685"} Feb 20 09:05:03 crc kubenswrapper[4948]: I0220 09:05:03.335684 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs7f6" event={"ID":"fb1bc246-7f5a-4060-8bdb-6a3963ea6862","Type":"ContainerStarted","Data":"a09d366d7a1bea20bacda46f1149340ebdb8bdbac35df3dd60174f0f350bbfa7"} Feb 20 09:05:03 crc kubenswrapper[4948]: I0220 09:05:03.473731 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:05:03 crc kubenswrapper[4948]: I0220 09:05:03.473776 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:05:03 crc kubenswrapper[4948]: I0220 09:05:03.534454 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:05:04 crc kubenswrapper[4948]: I0220 09:05:04.350881 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs7f6" event={"ID":"fb1bc246-7f5a-4060-8bdb-6a3963ea6862","Type":"ContainerStarted","Data":"d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85"} Feb 20 09:05:04 crc kubenswrapper[4948]: I0220 09:05:04.401709 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:05:05 crc kubenswrapper[4948]: I0220 09:05:05.365216 4948 generic.go:334] "Generic (PLEG): container finished" podID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerID="d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85" exitCode=0 Feb 20 09:05:05 crc kubenswrapper[4948]: I0220 09:05:05.365270 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs7f6" event={"ID":"fb1bc246-7f5a-4060-8bdb-6a3963ea6862","Type":"ContainerDied","Data":"d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85"} Feb 20 09:05:05 crc kubenswrapper[4948]: I0220 09:05:05.890705 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2hl84"] Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.378999 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs7f6" event={"ID":"fb1bc246-7f5a-4060-8bdb-6a3963ea6862","Type":"ContainerStarted","Data":"85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321"} Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.379163 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2hl84" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerName="registry-server" containerID="cri-o://42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca" gracePeriod=2 Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.406371 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gs7f6" podStartSLOduration=1.944916492 podStartE2EDuration="4.406350816s" podCreationTimestamp="2026-02-20 09:05:02 +0000 UTC" firstStartedPulling="2026-02-20 09:05:03.338386708 +0000 UTC m=+3552.312881528" lastFinishedPulling="2026-02-20 09:05:05.799821022 +0000 UTC m=+3554.774315852" observedRunningTime="2026-02-20 09:05:06.403062845 +0000 UTC m=+3555.377557665" watchObservedRunningTime="2026-02-20 09:05:06.406350816 +0000 UTC m=+3555.380845646" Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.852156 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.985332 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-catalog-content\") pod \"27f74a12-0843-4c61-a5c5-7f026224d86e\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.985468 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-utilities\") pod \"27f74a12-0843-4c61-a5c5-7f026224d86e\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.985590 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hv8x\" (UniqueName: \"kubernetes.io/projected/27f74a12-0843-4c61-a5c5-7f026224d86e-kube-api-access-6hv8x\") pod \"27f74a12-0843-4c61-a5c5-7f026224d86e\" (UID: \"27f74a12-0843-4c61-a5c5-7f026224d86e\") " Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.986910 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-utilities" (OuterVolumeSpecName: "utilities") pod "27f74a12-0843-4c61-a5c5-7f026224d86e" (UID: "27f74a12-0843-4c61-a5c5-7f026224d86e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:05:06 crc kubenswrapper[4948]: I0220 09:05:06.993345 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27f74a12-0843-4c61-a5c5-7f026224d86e-kube-api-access-6hv8x" (OuterVolumeSpecName: "kube-api-access-6hv8x") pod "27f74a12-0843-4c61-a5c5-7f026224d86e" (UID: "27f74a12-0843-4c61-a5c5-7f026224d86e"). InnerVolumeSpecName "kube-api-access-6hv8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.039192 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "27f74a12-0843-4c61-a5c5-7f026224d86e" (UID: "27f74a12-0843-4c61-a5c5-7f026224d86e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.087917 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.087958 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27f74a12-0843-4c61-a5c5-7f026224d86e-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.087984 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hv8x\" (UniqueName: \"kubernetes.io/projected/27f74a12-0843-4c61-a5c5-7f026224d86e-kube-api-access-6hv8x\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.389891 4948 generic.go:334] "Generic (PLEG): container finished" podID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerID="42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca" exitCode=0 Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.389992 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2hl84" event={"ID":"27f74a12-0843-4c61-a5c5-7f026224d86e","Type":"ContainerDied","Data":"42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca"} Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.390044 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2hl84" event={"ID":"27f74a12-0843-4c61-a5c5-7f026224d86e","Type":"ContainerDied","Data":"9d10453c1bde550a3b492982c9301c05c35c72c85021ea721f6d12fcfa8a9df5"} Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.390071 4948 scope.go:117] "RemoveContainer" containerID="42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.390230 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2hl84" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.433138 4948 scope.go:117] "RemoveContainer" containerID="b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.434456 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2hl84"] Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.444836 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2hl84"] Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.456858 4948 scope.go:117] "RemoveContainer" containerID="274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.496874 4948 scope.go:117] "RemoveContainer" containerID="42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca" Feb 20 09:05:07 crc kubenswrapper[4948]: E0220 09:05:07.497320 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca\": container with ID starting with 42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca not found: ID does not exist" containerID="42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.497406 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca"} err="failed to get container status \"42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca\": rpc error: code = NotFound desc = could not find container \"42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca\": container with ID starting with 42844cc459dfe5a86d4c469b9410a53b0ae7589280ea76862abd9dfba05593ca not found: ID does not exist" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.497471 4948 scope.go:117] "RemoveContainer" containerID="b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb" Feb 20 09:05:07 crc kubenswrapper[4948]: E0220 09:05:07.498056 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb\": container with ID starting with b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb not found: ID does not exist" containerID="b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.498125 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb"} err="failed to get container status \"b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb\": rpc error: code = NotFound desc = could not find container \"b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb\": container with ID starting with b940cc80e090937cbff77b5ed4366ab50ebcea5b7644547c7ab3fa22e20c4eeb not found: ID does not exist" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.498166 4948 scope.go:117] "RemoveContainer" containerID="274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76" Feb 20 09:05:07 crc kubenswrapper[4948]: E0220 09:05:07.498538 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76\": container with ID starting with 274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76 not found: ID does not exist" containerID="274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.498635 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76"} err="failed to get container status \"274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76\": rpc error: code = NotFound desc = could not find container \"274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76\": container with ID starting with 274f49b6cdbfb4b1df5dfe4a37883e06663d4ef43a70cc2e8f1bc96bbdbcfa76 not found: ID does not exist" Feb 20 09:05:07 crc kubenswrapper[4948]: I0220 09:05:07.734826 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" path="/var/lib/kubelet/pods/27f74a12-0843-4c61-a5c5-7f026224d86e/volumes" Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.025054 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.025500 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.025564 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.026565 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.026640 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" gracePeriod=600 Feb 20 09:05:08 crc kubenswrapper[4948]: E0220 09:05:08.169842 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.407250 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" exitCode=0 Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.407316 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2"} Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.407394 4948 scope.go:117] "RemoveContainer" containerID="b0e4da4a6a513a5c47c1f0dd87dcf0b56d3fede7e9008d839ac5623f61075ff8" Feb 20 09:05:08 crc kubenswrapper[4948]: I0220 09:05:08.408405 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:05:08 crc kubenswrapper[4948]: E0220 09:05:08.408881 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:05:11 crc kubenswrapper[4948]: I0220 09:05:11.498718 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8sqsd/must-gather-xzw72"] Feb 20 09:05:11 crc kubenswrapper[4948]: I0220 09:05:11.499308 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-8sqsd/must-gather-xzw72" podUID="b25c36d5-5de9-4c98-acf4-495d492def68" containerName="copy" containerID="cri-o://ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4" gracePeriod=2 Feb 20 09:05:11 crc kubenswrapper[4948]: I0220 09:05:11.520843 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8sqsd/must-gather-xzw72"] Feb 20 09:05:11 crc kubenswrapper[4948]: I0220 09:05:11.945543 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8sqsd_must-gather-xzw72_b25c36d5-5de9-4c98-acf4-495d492def68/copy/0.log" Feb 20 09:05:11 crc kubenswrapper[4948]: I0220 09:05:11.946344 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.024841 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsr8v\" (UniqueName: \"kubernetes.io/projected/b25c36d5-5de9-4c98-acf4-495d492def68-kube-api-access-vsr8v\") pod \"b25c36d5-5de9-4c98-acf4-495d492def68\" (UID: \"b25c36d5-5de9-4c98-acf4-495d492def68\") " Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.024929 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b25c36d5-5de9-4c98-acf4-495d492def68-must-gather-output\") pod \"b25c36d5-5de9-4c98-acf4-495d492def68\" (UID: \"b25c36d5-5de9-4c98-acf4-495d492def68\") " Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.030802 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b25c36d5-5de9-4c98-acf4-495d492def68-kube-api-access-vsr8v" (OuterVolumeSpecName: "kube-api-access-vsr8v") pod "b25c36d5-5de9-4c98-acf4-495d492def68" (UID: "b25c36d5-5de9-4c98-acf4-495d492def68"). InnerVolumeSpecName "kube-api-access-vsr8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.127529 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsr8v\" (UniqueName: \"kubernetes.io/projected/b25c36d5-5de9-4c98-acf4-495d492def68-kube-api-access-vsr8v\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.163937 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b25c36d5-5de9-4c98-acf4-495d492def68-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b25c36d5-5de9-4c98-acf4-495d492def68" (UID: "b25c36d5-5de9-4c98-acf4-495d492def68"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.229669 4948 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b25c36d5-5de9-4c98-acf4-495d492def68-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.428316 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.428397 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.492333 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8sqsd_must-gather-xzw72_b25c36d5-5de9-4c98-acf4-495d492def68/copy/0.log" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.493161 4948 generic.go:334] "Generic (PLEG): container finished" podID="b25c36d5-5de9-4c98-acf4-495d492def68" containerID="ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4" exitCode=143 Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.493261 4948 scope.go:117] "RemoveContainer" containerID="ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.493276 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8sqsd/must-gather-xzw72" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.530294 4948 scope.go:117] "RemoveContainer" containerID="4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.530403 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.599302 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.637938 4948 scope.go:117] "RemoveContainer" containerID="ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4" Feb 20 09:05:12 crc kubenswrapper[4948]: E0220 09:05:12.638592 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4\": container with ID starting with ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4 not found: ID does not exist" containerID="ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.638635 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4"} err="failed to get container status \"ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4\": rpc error: code = NotFound desc = could not find container \"ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4\": container with ID starting with ee095f93fada3e828226e407295d86f2e85f6d70e937ebcd69864d72a4cfb8b4 not found: ID does not exist" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.638660 4948 scope.go:117] "RemoveContainer" containerID="4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a" Feb 20 09:05:12 crc kubenswrapper[4948]: E0220 09:05:12.639163 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a\": container with ID starting with 4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a not found: ID does not exist" containerID="4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.639205 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a"} err="failed to get container status \"4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a\": rpc error: code = NotFound desc = could not find container \"4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a\": container with ID starting with 4a8285325687a203e4a7cd75d1ec55e566e63219bb19f28c257e627288bb520a not found: ID does not exist" Feb 20 09:05:12 crc kubenswrapper[4948]: I0220 09:05:12.782284 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gs7f6"] Feb 20 09:05:13 crc kubenswrapper[4948]: I0220 09:05:13.734795 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b25c36d5-5de9-4c98-acf4-495d492def68" path="/var/lib/kubelet/pods/b25c36d5-5de9-4c98-acf4-495d492def68/volumes" Feb 20 09:05:14 crc kubenswrapper[4948]: I0220 09:05:14.523124 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gs7f6" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerName="registry-server" containerID="cri-o://85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321" gracePeriod=2 Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.040628 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.086051 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-catalog-content\") pod \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.086186 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-utilities\") pod \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.086230 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmpwq\" (UniqueName: \"kubernetes.io/projected/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-kube-api-access-bmpwq\") pod \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\" (UID: \"fb1bc246-7f5a-4060-8bdb-6a3963ea6862\") " Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.088811 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-utilities" (OuterVolumeSpecName: "utilities") pod "fb1bc246-7f5a-4060-8bdb-6a3963ea6862" (UID: "fb1bc246-7f5a-4060-8bdb-6a3963ea6862"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.093184 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-kube-api-access-bmpwq" (OuterVolumeSpecName: "kube-api-access-bmpwq") pod "fb1bc246-7f5a-4060-8bdb-6a3963ea6862" (UID: "fb1bc246-7f5a-4060-8bdb-6a3963ea6862"). InnerVolumeSpecName "kube-api-access-bmpwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.155924 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fb1bc246-7f5a-4060-8bdb-6a3963ea6862" (UID: "fb1bc246-7f5a-4060-8bdb-6a3963ea6862"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.188794 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.188868 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.188881 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmpwq\" (UniqueName: \"kubernetes.io/projected/fb1bc246-7f5a-4060-8bdb-6a3963ea6862-kube-api-access-bmpwq\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.540246 4948 generic.go:334] "Generic (PLEG): container finished" podID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerID="85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321" exitCode=0 Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.540359 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs7f6" event={"ID":"fb1bc246-7f5a-4060-8bdb-6a3963ea6862","Type":"ContainerDied","Data":"85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321"} Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.540412 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs7f6" event={"ID":"fb1bc246-7f5a-4060-8bdb-6a3963ea6862","Type":"ContainerDied","Data":"a09d366d7a1bea20bacda46f1149340ebdb8bdbac35df3dd60174f0f350bbfa7"} Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.540452 4948 scope.go:117] "RemoveContainer" containerID="85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.540684 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gs7f6" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.572814 4948 scope.go:117] "RemoveContainer" containerID="d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.598033 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gs7f6"] Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.607230 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gs7f6"] Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.614680 4948 scope.go:117] "RemoveContainer" containerID="ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.665222 4948 scope.go:117] "RemoveContainer" containerID="85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321" Feb 20 09:05:15 crc kubenswrapper[4948]: E0220 09:05:15.665738 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321\": container with ID starting with 85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321 not found: ID does not exist" containerID="85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.665812 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321"} err="failed to get container status \"85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321\": rpc error: code = NotFound desc = could not find container \"85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321\": container with ID starting with 85b52de1041bebd75c7433b32671f81a45c491a3db70365771075aac31cce321 not found: ID does not exist" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.665852 4948 scope.go:117] "RemoveContainer" containerID="d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85" Feb 20 09:05:15 crc kubenswrapper[4948]: E0220 09:05:15.666494 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85\": container with ID starting with d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85 not found: ID does not exist" containerID="d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.666539 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85"} err="failed to get container status \"d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85\": rpc error: code = NotFound desc = could not find container \"d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85\": container with ID starting with d5289ac930f2ff05d94bb857ba6dd563c33915f3a369ab91433c852e800b4d85 not found: ID does not exist" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.666566 4948 scope.go:117] "RemoveContainer" containerID="ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685" Feb 20 09:05:15 crc kubenswrapper[4948]: E0220 09:05:15.667140 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685\": container with ID starting with ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685 not found: ID does not exist" containerID="ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.667202 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685"} err="failed to get container status \"ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685\": rpc error: code = NotFound desc = could not find container \"ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685\": container with ID starting with ac1cf50592cbedded2cb785f164d800b487c31663eb504bce695abf3f1e85685 not found: ID does not exist" Feb 20 09:05:15 crc kubenswrapper[4948]: I0220 09:05:15.745472 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" path="/var/lib/kubelet/pods/fb1bc246-7f5a-4060-8bdb-6a3963ea6862/volumes" Feb 20 09:05:23 crc kubenswrapper[4948]: I0220 09:05:23.723147 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:05:23 crc kubenswrapper[4948]: E0220 09:05:23.724078 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:05:38 crc kubenswrapper[4948]: I0220 09:05:38.723125 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:05:38 crc kubenswrapper[4948]: E0220 09:05:38.724399 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:05:51 crc kubenswrapper[4948]: I0220 09:05:51.723355 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:05:51 crc kubenswrapper[4948]: E0220 09:05:51.724440 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:06:02 crc kubenswrapper[4948]: I0220 09:06:02.723368 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:06:02 crc kubenswrapper[4948]: E0220 09:06:02.724527 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:06:03 crc kubenswrapper[4948]: I0220 09:06:03.021094 4948 scope.go:117] "RemoveContainer" containerID="d2d5efbac326ed9a4b4067346df3dcafa9289ae45429a1d4be737ea540693745" Feb 20 09:06:17 crc kubenswrapper[4948]: I0220 09:06:17.722857 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:06:17 crc kubenswrapper[4948]: E0220 09:06:17.730222 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:06:29 crc kubenswrapper[4948]: I0220 09:06:29.722600 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:06:29 crc kubenswrapper[4948]: E0220 09:06:29.723427 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:06:41 crc kubenswrapper[4948]: I0220 09:06:41.734804 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:06:41 crc kubenswrapper[4948]: E0220 09:06:41.735721 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:06:56 crc kubenswrapper[4948]: I0220 09:06:56.722855 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:06:56 crc kubenswrapper[4948]: E0220 09:06:56.724942 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:07:08 crc kubenswrapper[4948]: I0220 09:07:08.722401 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:07:08 crc kubenswrapper[4948]: E0220 09:07:08.723392 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:07:21 crc kubenswrapper[4948]: I0220 09:07:21.740918 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:07:21 crc kubenswrapper[4948]: E0220 09:07:21.742387 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:07:33 crc kubenswrapper[4948]: I0220 09:07:33.722607 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:07:33 crc kubenswrapper[4948]: E0220 09:07:33.723716 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:07:44 crc kubenswrapper[4948]: I0220 09:07:44.723088 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:07:44 crc kubenswrapper[4948]: E0220 09:07:44.723954 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:07:56 crc kubenswrapper[4948]: I0220 09:07:56.722757 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:07:56 crc kubenswrapper[4948]: E0220 09:07:56.723684 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:08:07 crc kubenswrapper[4948]: I0220 09:08:07.722692 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:08:07 crc kubenswrapper[4948]: E0220 09:08:07.723663 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.782239 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ql8t9/must-gather-hvzrp"] Feb 20 09:08:10 crc kubenswrapper[4948]: E0220 09:08:10.784444 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerName="registry-server" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.784571 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerName="registry-server" Feb 20 09:08:10 crc kubenswrapper[4948]: E0220 09:08:10.784677 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b25c36d5-5de9-4c98-acf4-495d492def68" containerName="copy" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.784768 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b25c36d5-5de9-4c98-acf4-495d492def68" containerName="copy" Feb 20 09:08:10 crc kubenswrapper[4948]: E0220 09:08:10.784891 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b25c36d5-5de9-4c98-acf4-495d492def68" containerName="gather" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.785002 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="b25c36d5-5de9-4c98-acf4-495d492def68" containerName="gather" Feb 20 09:08:10 crc kubenswrapper[4948]: E0220 09:08:10.785115 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerName="extract-utilities" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.785200 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerName="extract-utilities" Feb 20 09:08:10 crc kubenswrapper[4948]: E0220 09:08:10.785300 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerName="extract-content" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.785387 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerName="extract-content" Feb 20 09:08:10 crc kubenswrapper[4948]: E0220 09:08:10.785488 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerName="extract-content" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.785588 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerName="extract-content" Feb 20 09:08:10 crc kubenswrapper[4948]: E0220 09:08:10.785678 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerName="extract-utilities" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.785763 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerName="extract-utilities" Feb 20 09:08:10 crc kubenswrapper[4948]: E0220 09:08:10.785845 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerName="registry-server" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.785929 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerName="registry-server" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.786361 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b25c36d5-5de9-4c98-acf4-495d492def68" containerName="gather" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.786487 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb1bc246-7f5a-4060-8bdb-6a3963ea6862" containerName="registry-server" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.786633 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="27f74a12-0843-4c61-a5c5-7f026224d86e" containerName="registry-server" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.786738 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="b25c36d5-5de9-4c98-acf4-495d492def68" containerName="copy" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.788128 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.790162 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ql8t9/must-gather-hvzrp"] Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.790887 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ql8t9"/"kube-root-ca.crt" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.790980 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ql8t9"/"openshift-service-ca.crt" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.920335 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2bcm\" (UniqueName: \"kubernetes.io/projected/d81b349b-2494-4abc-b349-44fecd284709-kube-api-access-v2bcm\") pod \"must-gather-hvzrp\" (UID: \"d81b349b-2494-4abc-b349-44fecd284709\") " pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:08:10 crc kubenswrapper[4948]: I0220 09:08:10.920379 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d81b349b-2494-4abc-b349-44fecd284709-must-gather-output\") pod \"must-gather-hvzrp\" (UID: \"d81b349b-2494-4abc-b349-44fecd284709\") " pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:08:11 crc kubenswrapper[4948]: I0220 09:08:11.022503 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2bcm\" (UniqueName: \"kubernetes.io/projected/d81b349b-2494-4abc-b349-44fecd284709-kube-api-access-v2bcm\") pod \"must-gather-hvzrp\" (UID: \"d81b349b-2494-4abc-b349-44fecd284709\") " pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:08:11 crc kubenswrapper[4948]: I0220 09:08:11.022571 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d81b349b-2494-4abc-b349-44fecd284709-must-gather-output\") pod \"must-gather-hvzrp\" (UID: \"d81b349b-2494-4abc-b349-44fecd284709\") " pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:08:11 crc kubenswrapper[4948]: I0220 09:08:11.023303 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d81b349b-2494-4abc-b349-44fecd284709-must-gather-output\") pod \"must-gather-hvzrp\" (UID: \"d81b349b-2494-4abc-b349-44fecd284709\") " pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:08:11 crc kubenswrapper[4948]: I0220 09:08:11.049897 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2bcm\" (UniqueName: \"kubernetes.io/projected/d81b349b-2494-4abc-b349-44fecd284709-kube-api-access-v2bcm\") pod \"must-gather-hvzrp\" (UID: \"d81b349b-2494-4abc-b349-44fecd284709\") " pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:08:11 crc kubenswrapper[4948]: I0220 09:08:11.110308 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:08:11 crc kubenswrapper[4948]: I0220 09:08:11.549867 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ql8t9/must-gather-hvzrp"] Feb 20 09:08:11 crc kubenswrapper[4948]: W0220 09:08:11.550241 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd81b349b_2494_4abc_b349_44fecd284709.slice/crio-e4d5c7dd67fd11d407c5c350d2cbad2c4a79d8cf80019f91b10ec82e4e0a0226 WatchSource:0}: Error finding container e4d5c7dd67fd11d407c5c350d2cbad2c4a79d8cf80019f91b10ec82e4e0a0226: Status 404 returned error can't find the container with id e4d5c7dd67fd11d407c5c350d2cbad2c4a79d8cf80019f91b10ec82e4e0a0226 Feb 20 09:08:11 crc kubenswrapper[4948]: I0220 09:08:11.587683 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" event={"ID":"d81b349b-2494-4abc-b349-44fecd284709","Type":"ContainerStarted","Data":"e4d5c7dd67fd11d407c5c350d2cbad2c4a79d8cf80019f91b10ec82e4e0a0226"} Feb 20 09:08:12 crc kubenswrapper[4948]: I0220 09:08:12.599244 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" event={"ID":"d81b349b-2494-4abc-b349-44fecd284709","Type":"ContainerStarted","Data":"d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf"} Feb 20 09:08:12 crc kubenswrapper[4948]: I0220 09:08:12.599570 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" event={"ID":"d81b349b-2494-4abc-b349-44fecd284709","Type":"ContainerStarted","Data":"0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff"} Feb 20 09:08:12 crc kubenswrapper[4948]: I0220 09:08:12.622631 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" podStartSLOduration=2.622603205 podStartE2EDuration="2.622603205s" podCreationTimestamp="2026-02-20 09:08:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 09:08:12.616960336 +0000 UTC m=+3741.591455146" watchObservedRunningTime="2026-02-20 09:08:12.622603205 +0000 UTC m=+3741.597098045" Feb 20 09:08:15 crc kubenswrapper[4948]: E0220 09:08:15.150365 4948 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.113:47142->38.102.83.113:38489: write tcp 38.102.83.113:47142->38.102.83.113:38489: write: broken pipe Feb 20 09:08:15 crc kubenswrapper[4948]: I0220 09:08:15.825765 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-rdtwt"] Feb 20 09:08:15 crc kubenswrapper[4948]: I0220 09:08:15.826840 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:15 crc kubenswrapper[4948]: I0220 09:08:15.828777 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ql8t9"/"default-dockercfg-n7kqw" Feb 20 09:08:15 crc kubenswrapper[4948]: I0220 09:08:15.924603 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsf25\" (UniqueName: \"kubernetes.io/projected/25487e7a-c3d7-47c0-9ef0-1bd55109725a-kube-api-access-nsf25\") pod \"crc-debug-rdtwt\" (UID: \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\") " pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:15 crc kubenswrapper[4948]: I0220 09:08:15.924885 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25487e7a-c3d7-47c0-9ef0-1bd55109725a-host\") pod \"crc-debug-rdtwt\" (UID: \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\") " pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:16 crc kubenswrapper[4948]: I0220 09:08:16.027210 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsf25\" (UniqueName: \"kubernetes.io/projected/25487e7a-c3d7-47c0-9ef0-1bd55109725a-kube-api-access-nsf25\") pod \"crc-debug-rdtwt\" (UID: \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\") " pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:16 crc kubenswrapper[4948]: I0220 09:08:16.027841 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25487e7a-c3d7-47c0-9ef0-1bd55109725a-host\") pod \"crc-debug-rdtwt\" (UID: \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\") " pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:16 crc kubenswrapper[4948]: I0220 09:08:16.027954 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25487e7a-c3d7-47c0-9ef0-1bd55109725a-host\") pod \"crc-debug-rdtwt\" (UID: \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\") " pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:16 crc kubenswrapper[4948]: I0220 09:08:16.046302 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsf25\" (UniqueName: \"kubernetes.io/projected/25487e7a-c3d7-47c0-9ef0-1bd55109725a-kube-api-access-nsf25\") pod \"crc-debug-rdtwt\" (UID: \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\") " pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:16 crc kubenswrapper[4948]: I0220 09:08:16.143498 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:16 crc kubenswrapper[4948]: I0220 09:08:16.637731 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" event={"ID":"25487e7a-c3d7-47c0-9ef0-1bd55109725a","Type":"ContainerStarted","Data":"7079ff1f2f1f98bbc8f43f1d268a42718af20f413eed66d493a146a8c4b11eb9"} Feb 20 09:08:16 crc kubenswrapper[4948]: I0220 09:08:16.639155 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" event={"ID":"25487e7a-c3d7-47c0-9ef0-1bd55109725a","Type":"ContainerStarted","Data":"96045d0fbf4acb31ef0e8ddf75f287243e16367e34ffe86e106174591cc4fc32"} Feb 20 09:08:20 crc kubenswrapper[4948]: I0220 09:08:20.723202 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:08:20 crc kubenswrapper[4948]: E0220 09:08:20.723994 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:08:34 crc kubenswrapper[4948]: I0220 09:08:34.723818 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:08:34 crc kubenswrapper[4948]: E0220 09:08:34.724599 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:08:48 crc kubenswrapper[4948]: I0220 09:08:48.723798 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:08:48 crc kubenswrapper[4948]: E0220 09:08:48.725032 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:08:49 crc kubenswrapper[4948]: I0220 09:08:49.956827 4948 generic.go:334] "Generic (PLEG): container finished" podID="25487e7a-c3d7-47c0-9ef0-1bd55109725a" containerID="7079ff1f2f1f98bbc8f43f1d268a42718af20f413eed66d493a146a8c4b11eb9" exitCode=0 Feb 20 09:08:49 crc kubenswrapper[4948]: I0220 09:08:49.957230 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" event={"ID":"25487e7a-c3d7-47c0-9ef0-1bd55109725a","Type":"ContainerDied","Data":"7079ff1f2f1f98bbc8f43f1d268a42718af20f413eed66d493a146a8c4b11eb9"} Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.178300 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.218382 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-rdtwt"] Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.230525 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-rdtwt"] Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.303967 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsf25\" (UniqueName: \"kubernetes.io/projected/25487e7a-c3d7-47c0-9ef0-1bd55109725a-kube-api-access-nsf25\") pod \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\" (UID: \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\") " Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.304178 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25487e7a-c3d7-47c0-9ef0-1bd55109725a-host\") pod \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\" (UID: \"25487e7a-c3d7-47c0-9ef0-1bd55109725a\") " Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.304654 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/25487e7a-c3d7-47c0-9ef0-1bd55109725a-host" (OuterVolumeSpecName: "host") pod "25487e7a-c3d7-47c0-9ef0-1bd55109725a" (UID: "25487e7a-c3d7-47c0-9ef0-1bd55109725a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.309843 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25487e7a-c3d7-47c0-9ef0-1bd55109725a-kube-api-access-nsf25" (OuterVolumeSpecName: "kube-api-access-nsf25") pod "25487e7a-c3d7-47c0-9ef0-1bd55109725a" (UID: "25487e7a-c3d7-47c0-9ef0-1bd55109725a"). InnerVolumeSpecName "kube-api-access-nsf25". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.405908 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsf25\" (UniqueName: \"kubernetes.io/projected/25487e7a-c3d7-47c0-9ef0-1bd55109725a-kube-api-access-nsf25\") on node \"crc\" DevicePath \"\"" Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.405940 4948 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25487e7a-c3d7-47c0-9ef0-1bd55109725a-host\") on node \"crc\" DevicePath \"\"" Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.739041 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25487e7a-c3d7-47c0-9ef0-1bd55109725a" path="/var/lib/kubelet/pods/25487e7a-c3d7-47c0-9ef0-1bd55109725a/volumes" Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.971819 4948 scope.go:117] "RemoveContainer" containerID="7079ff1f2f1f98bbc8f43f1d268a42718af20f413eed66d493a146a8c4b11eb9" Feb 20 09:08:51 crc kubenswrapper[4948]: I0220 09:08:51.971853 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-rdtwt" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.393958 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-njlvs"] Feb 20 09:08:52 crc kubenswrapper[4948]: E0220 09:08:52.394305 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25487e7a-c3d7-47c0-9ef0-1bd55109725a" containerName="container-00" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.394317 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="25487e7a-c3d7-47c0-9ef0-1bd55109725a" containerName="container-00" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.394492 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="25487e7a-c3d7-47c0-9ef0-1bd55109725a" containerName="container-00" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.395098 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.398115 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ql8t9"/"default-dockercfg-n7kqw" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.525684 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-host\") pod \"crc-debug-njlvs\" (UID: \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\") " pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.525984 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjmgr\" (UniqueName: \"kubernetes.io/projected/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-kube-api-access-cjmgr\") pod \"crc-debug-njlvs\" (UID: \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\") " pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.627553 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-host\") pod \"crc-debug-njlvs\" (UID: \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\") " pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.627670 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjmgr\" (UniqueName: \"kubernetes.io/projected/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-kube-api-access-cjmgr\") pod \"crc-debug-njlvs\" (UID: \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\") " pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.627769 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-host\") pod \"crc-debug-njlvs\" (UID: \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\") " pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.648154 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjmgr\" (UniqueName: \"kubernetes.io/projected/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-kube-api-access-cjmgr\") pod \"crc-debug-njlvs\" (UID: \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\") " pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.726609 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:52 crc kubenswrapper[4948]: W0220 09:08:52.789023 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod946347cb_7ccd_45e9_866c_2f9b2d6e10cb.slice/crio-dc66a6b6c2a4cedaf7ccafd526de87fb780940e37aeeeb99e15e84f58fc897f3 WatchSource:0}: Error finding container dc66a6b6c2a4cedaf7ccafd526de87fb780940e37aeeeb99e15e84f58fc897f3: Status 404 returned error can't find the container with id dc66a6b6c2a4cedaf7ccafd526de87fb780940e37aeeeb99e15e84f58fc897f3 Feb 20 09:08:52 crc kubenswrapper[4948]: I0220 09:08:52.981573 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/crc-debug-njlvs" event={"ID":"946347cb-7ccd-45e9-866c-2f9b2d6e10cb","Type":"ContainerStarted","Data":"dc66a6b6c2a4cedaf7ccafd526de87fb780940e37aeeeb99e15e84f58fc897f3"} Feb 20 09:08:53 crc kubenswrapper[4948]: I0220 09:08:53.992356 4948 generic.go:334] "Generic (PLEG): container finished" podID="946347cb-7ccd-45e9-866c-2f9b2d6e10cb" containerID="01b0492d26413e240c32017d4eb44b4ec84b1e4c3a3eedc722410b0fbfdd91a9" exitCode=0 Feb 20 09:08:53 crc kubenswrapper[4948]: I0220 09:08:53.992403 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/crc-debug-njlvs" event={"ID":"946347cb-7ccd-45e9-866c-2f9b2d6e10cb","Type":"ContainerDied","Data":"01b0492d26413e240c32017d4eb44b4ec84b1e4c3a3eedc722410b0fbfdd91a9"} Feb 20 09:08:54 crc kubenswrapper[4948]: I0220 09:08:54.428870 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-njlvs"] Feb 20 09:08:54 crc kubenswrapper[4948]: I0220 09:08:54.440280 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-njlvs"] Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.088411 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.171699 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjmgr\" (UniqueName: \"kubernetes.io/projected/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-kube-api-access-cjmgr\") pod \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\" (UID: \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\") " Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.171904 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-host\") pod \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\" (UID: \"946347cb-7ccd-45e9-866c-2f9b2d6e10cb\") " Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.172049 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-host" (OuterVolumeSpecName: "host") pod "946347cb-7ccd-45e9-866c-2f9b2d6e10cb" (UID: "946347cb-7ccd-45e9-866c-2f9b2d6e10cb"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.172354 4948 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-host\") on node \"crc\" DevicePath \"\"" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.184722 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-kube-api-access-cjmgr" (OuterVolumeSpecName: "kube-api-access-cjmgr") pod "946347cb-7ccd-45e9-866c-2f9b2d6e10cb" (UID: "946347cb-7ccd-45e9-866c-2f9b2d6e10cb"). InnerVolumeSpecName "kube-api-access-cjmgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.274499 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjmgr\" (UniqueName: \"kubernetes.io/projected/946347cb-7ccd-45e9-866c-2f9b2d6e10cb-kube-api-access-cjmgr\") on node \"crc\" DevicePath \"\"" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.657787 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-5gtxj"] Feb 20 09:08:55 crc kubenswrapper[4948]: E0220 09:08:55.658171 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946347cb-7ccd-45e9-866c-2f9b2d6e10cb" containerName="container-00" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.658189 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="946347cb-7ccd-45e9-866c-2f9b2d6e10cb" containerName="container-00" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.658367 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="946347cb-7ccd-45e9-866c-2f9b2d6e10cb" containerName="container-00" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.658950 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.732094 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="946347cb-7ccd-45e9-866c-2f9b2d6e10cb" path="/var/lib/kubelet/pods/946347cb-7ccd-45e9-866c-2f9b2d6e10cb/volumes" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.785412 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t82dt\" (UniqueName: \"kubernetes.io/projected/0cfde212-7150-4b57-b3dd-41715c736426-kube-api-access-t82dt\") pod \"crc-debug-5gtxj\" (UID: \"0cfde212-7150-4b57-b3dd-41715c736426\") " pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.785725 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfde212-7150-4b57-b3dd-41715c736426-host\") pod \"crc-debug-5gtxj\" (UID: \"0cfde212-7150-4b57-b3dd-41715c736426\") " pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.887907 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t82dt\" (UniqueName: \"kubernetes.io/projected/0cfde212-7150-4b57-b3dd-41715c736426-kube-api-access-t82dt\") pod \"crc-debug-5gtxj\" (UID: \"0cfde212-7150-4b57-b3dd-41715c736426\") " pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.888612 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfde212-7150-4b57-b3dd-41715c736426-host\") pod \"crc-debug-5gtxj\" (UID: \"0cfde212-7150-4b57-b3dd-41715c736426\") " pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.888443 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfde212-7150-4b57-b3dd-41715c736426-host\") pod \"crc-debug-5gtxj\" (UID: \"0cfde212-7150-4b57-b3dd-41715c736426\") " pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.905889 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t82dt\" (UniqueName: \"kubernetes.io/projected/0cfde212-7150-4b57-b3dd-41715c736426-kube-api-access-t82dt\") pod \"crc-debug-5gtxj\" (UID: \"0cfde212-7150-4b57-b3dd-41715c736426\") " pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:55 crc kubenswrapper[4948]: I0220 09:08:55.974357 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:56 crc kubenswrapper[4948]: I0220 09:08:56.030173 4948 scope.go:117] "RemoveContainer" containerID="01b0492d26413e240c32017d4eb44b4ec84b1e4c3a3eedc722410b0fbfdd91a9" Feb 20 09:08:56 crc kubenswrapper[4948]: I0220 09:08:56.030308 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-njlvs" Feb 20 09:08:56 crc kubenswrapper[4948]: I0220 09:08:56.042116 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" event={"ID":"0cfde212-7150-4b57-b3dd-41715c736426","Type":"ContainerStarted","Data":"6f9248e9b3243b7c97c4714220f5bdb0e400fa6c3ef7be9e1e20648a82dc1da6"} Feb 20 09:08:57 crc kubenswrapper[4948]: I0220 09:08:57.054345 4948 generic.go:334] "Generic (PLEG): container finished" podID="0cfde212-7150-4b57-b3dd-41715c736426" containerID="cf4963aa778b92fd84cff1d5614e4d02e1a3608c465ada5c0b7912612f7613de" exitCode=0 Feb 20 09:08:57 crc kubenswrapper[4948]: I0220 09:08:57.054418 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" event={"ID":"0cfde212-7150-4b57-b3dd-41715c736426","Type":"ContainerDied","Data":"cf4963aa778b92fd84cff1d5614e4d02e1a3608c465ada5c0b7912612f7613de"} Feb 20 09:08:57 crc kubenswrapper[4948]: I0220 09:08:57.107942 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-5gtxj"] Feb 20 09:08:57 crc kubenswrapper[4948]: I0220 09:08:57.117287 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ql8t9/crc-debug-5gtxj"] Feb 20 09:08:58 crc kubenswrapper[4948]: I0220 09:08:58.192697 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:58 crc kubenswrapper[4948]: I0220 09:08:58.339256 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t82dt\" (UniqueName: \"kubernetes.io/projected/0cfde212-7150-4b57-b3dd-41715c736426-kube-api-access-t82dt\") pod \"0cfde212-7150-4b57-b3dd-41715c736426\" (UID: \"0cfde212-7150-4b57-b3dd-41715c736426\") " Feb 20 09:08:58 crc kubenswrapper[4948]: I0220 09:08:58.339430 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfde212-7150-4b57-b3dd-41715c736426-host\") pod \"0cfde212-7150-4b57-b3dd-41715c736426\" (UID: \"0cfde212-7150-4b57-b3dd-41715c736426\") " Feb 20 09:08:58 crc kubenswrapper[4948]: I0220 09:08:58.339563 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfde212-7150-4b57-b3dd-41715c736426-host" (OuterVolumeSpecName: "host") pod "0cfde212-7150-4b57-b3dd-41715c736426" (UID: "0cfde212-7150-4b57-b3dd-41715c736426"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 09:08:58 crc kubenswrapper[4948]: I0220 09:08:58.339871 4948 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfde212-7150-4b57-b3dd-41715c736426-host\") on node \"crc\" DevicePath \"\"" Feb 20 09:08:58 crc kubenswrapper[4948]: I0220 09:08:58.344231 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cfde212-7150-4b57-b3dd-41715c736426-kube-api-access-t82dt" (OuterVolumeSpecName: "kube-api-access-t82dt") pod "0cfde212-7150-4b57-b3dd-41715c736426" (UID: "0cfde212-7150-4b57-b3dd-41715c736426"). InnerVolumeSpecName "kube-api-access-t82dt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:08:58 crc kubenswrapper[4948]: I0220 09:08:58.441401 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t82dt\" (UniqueName: \"kubernetes.io/projected/0cfde212-7150-4b57-b3dd-41715c736426-kube-api-access-t82dt\") on node \"crc\" DevicePath \"\"" Feb 20 09:08:59 crc kubenswrapper[4948]: I0220 09:08:59.078070 4948 scope.go:117] "RemoveContainer" containerID="cf4963aa778b92fd84cff1d5614e4d02e1a3608c465ada5c0b7912612f7613de" Feb 20 09:08:59 crc kubenswrapper[4948]: I0220 09:08:59.078351 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/crc-debug-5gtxj" Feb 20 09:08:59 crc kubenswrapper[4948]: I0220 09:08:59.722235 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:08:59 crc kubenswrapper[4948]: E0220 09:08:59.722606 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:08:59 crc kubenswrapper[4948]: I0220 09:08:59.741744 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cfde212-7150-4b57-b3dd-41715c736426" path="/var/lib/kubelet/pods/0cfde212-7150-4b57-b3dd-41715c736426/volumes" Feb 20 09:09:11 crc kubenswrapper[4948]: I0220 09:09:11.728218 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:09:11 crc kubenswrapper[4948]: E0220 09:09:11.728940 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:09:24 crc kubenswrapper[4948]: I0220 09:09:24.723393 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:09:24 crc kubenswrapper[4948]: E0220 09:09:24.724370 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:09:30 crc kubenswrapper[4948]: I0220 09:09:30.743588 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-686d666564-8tnjb_51f6301c-57be-4a07-91df-b1a9f90e3f40/barbican-api/0.log" Feb 20 09:09:30 crc kubenswrapper[4948]: I0220 09:09:30.880665 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-686d666564-8tnjb_51f6301c-57be-4a07-91df-b1a9f90e3f40/barbican-api-log/0.log" Feb 20 09:09:30 crc kubenswrapper[4948]: I0220 09:09:30.939532 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6ffd468-xgk27_292d8e1e-de3b-4259-afef-c19e2c21187c/barbican-keystone-listener/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.026226 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6ffd468-xgk27_292d8e1e-de3b-4259-afef-c19e2c21187c/barbican-keystone-listener-log/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.122477 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76d486c65f-w9lc6_4b356058-d957-4fa5-a106-3fe5052c8539/barbican-worker/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.187164 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76d486c65f-w9lc6_4b356058-d957-4fa5-a106-3fe5052c8539/barbican-worker-log/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.314606 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-4265g_e9bed94f-dc85-433b-b2db-8da400959f54/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.416195 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58f563bb-9716-4abd-b187-3f01036d4f31/ceilometer-central-agent/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.450624 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58f563bb-9716-4abd-b187-3f01036d4f31/ceilometer-notification-agent/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.484522 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58f563bb-9716-4abd-b187-3f01036d4f31/proxy-httpd/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.578693 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58f563bb-9716-4abd-b187-3f01036d4f31/sg-core/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.677176 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e5967cab-cc29-4c87-b873-0f5d20203d4f/cinder-api-log/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.682705 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e5967cab-cc29-4c87-b873-0f5d20203d4f/cinder-api/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.890280 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2f8a0f7b-ee05-4e93-a591-7486fecf8cd0/cinder-scheduler/0.log" Feb 20 09:09:31 crc kubenswrapper[4948]: I0220 09:09:31.893314 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2f8a0f7b-ee05-4e93-a591-7486fecf8cd0/probe/0.log" Feb 20 09:09:32 crc kubenswrapper[4948]: I0220 09:09:32.095689 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-mz5vx_bf3501bf-8f1e-4529-91a8-6be83eda4158/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:32 crc kubenswrapper[4948]: I0220 09:09:32.098565 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-5qbq6_d38195a4-0c7b-427e-b5e2-923b24c10674/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:32 crc kubenswrapper[4948]: I0220 09:09:32.419807 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-fbjpk_b05f92c6-60c9-41be-bf6c-6288cee04659/init/0.log" Feb 20 09:09:32 crc kubenswrapper[4948]: I0220 09:09:32.614023 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-fbjpk_b05f92c6-60c9-41be-bf6c-6288cee04659/init/0.log" Feb 20 09:09:32 crc kubenswrapper[4948]: I0220 09:09:32.699614 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-mhkll_5864a68d-650e-4bcf-b705-619c0f27445b/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:32 crc kubenswrapper[4948]: I0220 09:09:32.702710 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-fbjpk_b05f92c6-60c9-41be-bf6c-6288cee04659/dnsmasq-dns/0.log" Feb 20 09:09:32 crc kubenswrapper[4948]: I0220 09:09:32.920952 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a/glance-log/0.log" Feb 20 09:09:32 crc kubenswrapper[4948]: I0220 09:09:32.970139 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_acd5bc1f-fee5-4ef1-95f9-5c4d11bbc54a/glance-httpd/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.059730 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4fcb965e-4d04-4863-a966-39a83f458fa6/glance-httpd/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.145268 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4fcb965e-4d04-4863-a966-39a83f458fa6/glance-log/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.242401 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-685c4db87b-rw57h_7b9a072a-a040-414c-96c7-cf1454148745/horizon/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.443002 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-8fgpw_c517042a-5354-4a8d-b7ff-c9aafa263b6c/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.554206 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-685c4db87b-rw57h_7b9a072a-a040-414c-96c7-cf1454148745/horizon-log/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.582627 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-4l55v_f42ddd0c-a2ec-414a-812d-f87213a3226d/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.745039 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29526301-gklw9_7990f62f-cd0f-4be2-b166-c00621d877e1/keystone-cron/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.859656 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6d4b8df464-wspg2_6687eafb-7d21-45d0-8ff7-f3c1a1d4d3bc/keystone-api/0.log" Feb 20 09:09:33 crc kubenswrapper[4948]: I0220 09:09:33.898533 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_24bf0fb4-81db-4955-8b5f-92c0f6b0856b/kube-state-metrics/0.log" Feb 20 09:09:34 crc kubenswrapper[4948]: I0220 09:09:34.062616 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-nmwvs_6e619eac-dfdc-4c8c-aa51-fc08ba58b49a/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:34 crc kubenswrapper[4948]: I0220 09:09:34.381364 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-76b5684765-h5624_fb9d5a3e-b07f-4229-9b5b-d0db6750cf59/neutron-httpd/0.log" Feb 20 09:09:34 crc kubenswrapper[4948]: I0220 09:09:34.414673 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-76b5684765-h5624_fb9d5a3e-b07f-4229-9b5b-d0db6750cf59/neutron-api/0.log" Feb 20 09:09:34 crc kubenswrapper[4948]: I0220 09:09:34.472473 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-q5zcr_fd280342-4276-4eee-a763-876e31be28c5/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:34 crc kubenswrapper[4948]: I0220 09:09:34.956432 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_354f9b2f-9eaa-4e47-b214-0fd895f90c4c/nova-api-log/0.log" Feb 20 09:09:35 crc kubenswrapper[4948]: I0220 09:09:35.175688 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_873a6553-9637-4f2a-a743-f33cde7fc883/nova-cell0-conductor-conductor/0.log" Feb 20 09:09:35 crc kubenswrapper[4948]: I0220 09:09:35.411454 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_e801ada9-0247-4b29-b262-04637e1f8452/nova-cell1-conductor-conductor/0.log" Feb 20 09:09:35 crc kubenswrapper[4948]: I0220 09:09:35.519498 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9c00f24b-fedb-4747-91c4-3d0b551a7288/nova-cell1-novncproxy-novncproxy/0.log" Feb 20 09:09:35 crc kubenswrapper[4948]: I0220 09:09:35.531564 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_354f9b2f-9eaa-4e47-b214-0fd895f90c4c/nova-api-api/0.log" Feb 20 09:09:35 crc kubenswrapper[4948]: I0220 09:09:35.723683 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:09:35 crc kubenswrapper[4948]: E0220 09:09:35.723993 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:09:35 crc kubenswrapper[4948]: I0220 09:09:35.843604 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-p22xr_6005ded0-9e21-4908-bbf0-33c710ba4341/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:36 crc kubenswrapper[4948]: I0220 09:09:36.083005 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_59965363-9971-4130-bb5a-5fcf5a44e2d5/nova-metadata-log/0.log" Feb 20 09:09:36 crc kubenswrapper[4948]: I0220 09:09:36.341176 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d02114ae-21f2-4cc5-9bca-9505c25eaf5f/mysql-bootstrap/0.log" Feb 20 09:09:36 crc kubenswrapper[4948]: I0220 09:09:36.385566 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_f114bc41-01a5-4955-97eb-7fcf139cc5a9/nova-scheduler-scheduler/0.log" Feb 20 09:09:36 crc kubenswrapper[4948]: I0220 09:09:36.509923 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d02114ae-21f2-4cc5-9bca-9505c25eaf5f/mysql-bootstrap/0.log" Feb 20 09:09:36 crc kubenswrapper[4948]: I0220 09:09:36.539391 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d02114ae-21f2-4cc5-9bca-9505c25eaf5f/galera/0.log" Feb 20 09:09:36 crc kubenswrapper[4948]: I0220 09:09:36.757428 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e59b63d-0c56-488c-87cb-348af87058c6/mysql-bootstrap/0.log" Feb 20 09:09:36 crc kubenswrapper[4948]: I0220 09:09:36.957406 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e59b63d-0c56-488c-87cb-348af87058c6/mysql-bootstrap/0.log" Feb 20 09:09:36 crc kubenswrapper[4948]: I0220 09:09:36.975160 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6e59b63d-0c56-488c-87cb-348af87058c6/galera/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.128538 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_b2a76f68-3b4b-4993-aa4f-3b31b4ee1008/openstackclient/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.259184 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-gcfn8_1510e8b2-af6f-4641-a2d2-361830382250/openstack-network-exporter/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.301333 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_59965363-9971-4130-bb5a-5fcf5a44e2d5/nova-metadata-metadata/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.477497 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-mkpng_b80cb988-de2a-4e65-9161-a0af0561c754/ovn-controller/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.501379 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q4bv4_c82772b6-9737-45ee-998a-e897086c03b5/ovsdb-server-init/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.691377 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q4bv4_c82772b6-9737-45ee-998a-e897086c03b5/ovsdb-server-init/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.694023 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q4bv4_c82772b6-9737-45ee-998a-e897086c03b5/ovs-vswitchd/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.710111 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q4bv4_c82772b6-9737-45ee-998a-e897086c03b5/ovsdb-server/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.950432 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-4b2vp_b19ca7cb-01d2-4965-90f6-5d50211870cc/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:37 crc kubenswrapper[4948]: I0220 09:09:37.988789 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_eeb3782c-0a88-4d17-b481-e2aef7ea386a/openstack-network-exporter/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.127300 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_eeb3782c-0a88-4d17-b481-e2aef7ea386a/ovn-northd/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.153530 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d681442e-6bdb-46e4-9b49-6466e5e036a0/openstack-network-exporter/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.193936 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d681442e-6bdb-46e4-9b49-6466e5e036a0/ovsdbserver-nb/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.344135 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4b04e19d-cd3d-474d-884d-df12d0e1fffd/openstack-network-exporter/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.450488 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4b04e19d-cd3d-474d-884d-df12d0e1fffd/ovsdbserver-sb/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.577387 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-86887846d-pqjcb_5047ff00-845b-459f-ac89-6cbf997f93fb/placement-api/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.662231 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-86887846d-pqjcb_5047ff00-845b-459f-ac89-6cbf997f93fb/placement-log/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.666536 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_037b329f-712f-4a67-984a-75affd2a57b7/setup-container/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.954844 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_037b329f-712f-4a67-984a-75affd2a57b7/setup-container/0.log" Feb 20 09:09:38 crc kubenswrapper[4948]: I0220 09:09:38.986918 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_037b329f-712f-4a67-984a-75affd2a57b7/rabbitmq/0.log" Feb 20 09:09:39 crc kubenswrapper[4948]: I0220 09:09:39.173303 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f063bb62-10c3-46d9-a41d-56d2826059e0/setup-container/0.log" Feb 20 09:09:39 crc kubenswrapper[4948]: I0220 09:09:39.300011 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f063bb62-10c3-46d9-a41d-56d2826059e0/setup-container/0.log" Feb 20 09:09:39 crc kubenswrapper[4948]: I0220 09:09:39.391738 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f063bb62-10c3-46d9-a41d-56d2826059e0/rabbitmq/0.log" Feb 20 09:09:39 crc kubenswrapper[4948]: I0220 09:09:39.485199 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-ds522_e3b05342-f970-4fdb-9b1c-ff521e118f96/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:39 crc kubenswrapper[4948]: I0220 09:09:39.568728 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-f7ww9_c3e7cce4-5144-4190-8013-83428c66cde9/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:39 crc kubenswrapper[4948]: I0220 09:09:39.700632 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-fjbch_90b6e7df-b3e1-4d05-a6b3-7fe47d70804f/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:39 crc kubenswrapper[4948]: I0220 09:09:39.800522 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-lkbkf_79aadb90-7625-4eb5-8cd9-e9a3e7e6afb8/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.009594 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-cqchc_41856163-b621-439a-95c1-ca8d13f0a08c/ssh-known-hosts-edpm-deployment/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.172553 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7c7cd9876c-262qf_30f48966-e44d-44bc-a868-7e8119ac186c/proxy-httpd/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.225120 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-dllls_642adfd1-e253-4e0a-80e6-860a9f5c1ae1/swift-ring-rebalance/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.228202 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7c7cd9876c-262qf_30f48966-e44d-44bc-a868-7e8119ac186c/proxy-server/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.410742 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/account-reaper/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.420955 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/account-auditor/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.467643 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/account-replicator/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.608333 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/container-auditor/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.612558 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/account-server/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.672175 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/container-server/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.674237 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/container-replicator/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.783919 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/container-updater/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.816808 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-auditor/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.898573 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-expirer/0.log" Feb 20 09:09:40 crc kubenswrapper[4948]: I0220 09:09:40.908554 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-replicator/0.log" Feb 20 09:09:41 crc kubenswrapper[4948]: I0220 09:09:41.021593 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-updater/0.log" Feb 20 09:09:41 crc kubenswrapper[4948]: I0220 09:09:41.026903 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/object-server/0.log" Feb 20 09:09:41 crc kubenswrapper[4948]: I0220 09:09:41.101100 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/rsync/0.log" Feb 20 09:09:41 crc kubenswrapper[4948]: I0220 09:09:41.106834 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e4d60def-5dc9-460b-92ed-e3d0157982ec/swift-recon-cron/0.log" Feb 20 09:09:41 crc kubenswrapper[4948]: I0220 09:09:41.339892 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-nmlxd_06ed5839-bc31-4691-8586-cd42c0413006/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:41 crc kubenswrapper[4948]: I0220 09:09:41.358128 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_0fefa0c4-ece9-4caf-8b41-8a64472c53e6/tempest-tests-tempest-tests-runner/0.log" Feb 20 09:09:41 crc kubenswrapper[4948]: I0220 09:09:41.507949 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_de083766-efe0-42cd-95c5-b87bbeb33b0a/test-operator-logs-container/0.log" Feb 20 09:09:41 crc kubenswrapper[4948]: I0220 09:09:41.611710 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-lf98d_2e05be38-f4ab-415f-a71d-9e5233cf0ea7/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:09:50 crc kubenswrapper[4948]: I0220 09:09:50.722192 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:09:50 crc kubenswrapper[4948]: E0220 09:09:50.723908 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:09:52 crc kubenswrapper[4948]: I0220 09:09:52.316865 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_8e994489-a2cf-4f6d-a00c-98f627ba0e5f/memcached/0.log" Feb 20 09:10:04 crc kubenswrapper[4948]: I0220 09:10:04.722494 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:10:04 crc kubenswrapper[4948]: E0220 09:10:04.723373 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:10:07 crc kubenswrapper[4948]: I0220 09:10:07.307269 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/util/0.log" Feb 20 09:10:07 crc kubenswrapper[4948]: I0220 09:10:07.510607 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/pull/0.log" Feb 20 09:10:07 crc kubenswrapper[4948]: I0220 09:10:07.514336 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/util/0.log" Feb 20 09:10:07 crc kubenswrapper[4948]: I0220 09:10:07.529193 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/pull/0.log" Feb 20 09:10:07 crc kubenswrapper[4948]: I0220 09:10:07.694107 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/extract/0.log" Feb 20 09:10:07 crc kubenswrapper[4948]: I0220 09:10:07.714485 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/pull/0.log" Feb 20 09:10:07 crc kubenswrapper[4948]: I0220 09:10:07.719090 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b02772c0e8c97b926926e0a1d8d8c995d8f01d9d9d64402b28cb4393dfnlxt6_a1b99eb4-2ab2-4755-92ed-a5226d91bfaf/util/0.log" Feb 20 09:10:08 crc kubenswrapper[4948]: I0220 09:10:08.232588 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-kjpkr_98a1aa06-948b-4034-bc07-7e546e341a8f/manager/0.log" Feb 20 09:10:08 crc kubenswrapper[4948]: I0220 09:10:08.492153 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987464f4-xc74s_347cdd7c-e5e8-49b1-a4f1-687b8a06b250/manager/0.log" Feb 20 09:10:08 crc kubenswrapper[4948]: I0220 09:10:08.659516 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-6cr5p_2d303d1a-1062-4bf6-be68-1bd6d1a3228f/manager/0.log" Feb 20 09:10:08 crc kubenswrapper[4948]: I0220 09:10:08.901883 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-bkl49_4a663826-d4b5-4ed3-8270-099b003390b4/manager/0.log" Feb 20 09:10:09 crc kubenswrapper[4948]: I0220 09:10:09.192436 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5d946d989d-c6xqd_586113f2-38b6-4bd1-8adb-3c155bb35ba9/manager/0.log" Feb 20 09:10:09 crc kubenswrapper[4948]: I0220 09:10:09.366201 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-5fqzw_105a118c-121d-4582-960f-1da9957980cb/manager/0.log" Feb 20 09:10:09 crc kubenswrapper[4948]: I0220 09:10:09.414728 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79d975b745-jrlw5_93b93501-acfe-4274-9a22-ca644b1d11d3/manager/0.log" Feb 20 09:10:09 crc kubenswrapper[4948]: I0220 09:10:09.664120 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-ttkfm_3fbbd11d-612d-479a-a34f-505d995a4871/manager/0.log" Feb 20 09:10:09 crc kubenswrapper[4948]: I0220 09:10:09.667161 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-54f6768c69-lj2sz_f762ec0d-e09c-4baf-9540-f4ac61ce7234/manager/0.log" Feb 20 09:10:09 crc kubenswrapper[4948]: I0220 09:10:09.851923 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-zcccd_6e8e8b87-5a00-486f-a00c-1450c109f3b2/manager/0.log" Feb 20 09:10:10 crc kubenswrapper[4948]: I0220 09:10:10.388929 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64ddbf8bb-jcnxd_88142137-864d-4660-a688-a7dcc503851b/manager/0.log" Feb 20 09:10:10 crc kubenswrapper[4948]: I0220 09:10:10.456535 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-jxwsx_7c7bb531-7900-4cc6-9d9b-bae52dabc59d/manager/0.log" Feb 20 09:10:10 crc kubenswrapper[4948]: I0220 09:10:10.699524 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c6767dc9cdh46m_b8b07685-095c-4bbd-a30d-57ca59d7cbdc/manager/0.log" Feb 20 09:10:11 crc kubenswrapper[4948]: I0220 09:10:11.021034 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-59c446d8d6-4cs2n_9465a7a8-c08e-44c9-a76d-3878b0bff3af/operator/0.log" Feb 20 09:10:11 crc kubenswrapper[4948]: I0220 09:10:11.306257 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-9z4jj_64872326-fb95-4bc1-a6c5-0b34242883f9/registry-server/0.log" Feb 20 09:10:11 crc kubenswrapper[4948]: I0220 09:10:11.560285 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-d44cf6b75-dkkq9_3ea675de-d1b0-4880-9652-eb066f6b0fb7/manager/0.log" Feb 20 09:10:11 crc kubenswrapper[4948]: I0220 09:10:11.799746 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-znrkz_16933747-642c-45ff-9f98-9321c633826a/manager/0.log" Feb 20 09:10:11 crc kubenswrapper[4948]: I0220 09:10:11.946899 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f8888797-hl5w7_ee33bd02-a3ce-415d-9d2d-fefd383e9810/manager/0.log" Feb 20 09:10:12 crc kubenswrapper[4948]: I0220 09:10:12.004503 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-59pcs_fe0aca33-1acf-463a-91a6-bbf35a38fd7c/operator/0.log" Feb 20 09:10:12 crc kubenswrapper[4948]: I0220 09:10:12.183362 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-5lfts_d65e0993-11a0-4e81-963f-eeb9dcb92536/manager/0.log" Feb 20 09:10:12 crc kubenswrapper[4948]: I0220 09:10:12.297104 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f45b4ff68-5rrd9_0f0cbb65-3c8e-41e6-8059-34e121de0821/manager/0.log" Feb 20 09:10:12 crc kubenswrapper[4948]: I0220 09:10:12.363796 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7866795846-f7cxn_0f98bc58-d0e3-405b-88fd-d8bd65f415a4/manager/0.log" Feb 20 09:10:12 crc kubenswrapper[4948]: I0220 09:10:12.577549 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5db88f68c-pndq7_f2400173-aa87-476c-8216-4f8c9cf9d474/manager/0.log" Feb 20 09:10:12 crc kubenswrapper[4948]: I0220 09:10:12.808701 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8569bc6fdb-lt4jk_e9f13bcd-ac5a-4cbd-952a-a5b9dceee562/manager/0.log" Feb 20 09:10:14 crc kubenswrapper[4948]: I0220 09:10:14.443678 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-4h9v5_b8e12d0c-9564-4bf2-ac61-b22d2fbdf855/manager/0.log" Feb 20 09:10:19 crc kubenswrapper[4948]: I0220 09:10:19.723158 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:10:20 crc kubenswrapper[4948]: I0220 09:10:20.814970 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"5bbfc48505968074b16a56aea08c035f5c0687419ea24b66d69a44ebd2ab9d7d"} Feb 20 09:10:31 crc kubenswrapper[4948]: I0220 09:10:31.561016 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-bgm85_a149fe37-c748-4120-9116-1da4b680d880/control-plane-machine-set-operator/0.log" Feb 20 09:10:31 crc kubenswrapper[4948]: I0220 09:10:31.717782 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fthdq_5bc3d0cd-eacd-4b25-9acd-853e49db7b47/kube-rbac-proxy/0.log" Feb 20 09:10:31 crc kubenswrapper[4948]: I0220 09:10:31.718723 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fthdq_5bc3d0cd-eacd-4b25-9acd-853e49db7b47/machine-api-operator/0.log" Feb 20 09:10:45 crc kubenswrapper[4948]: I0220 09:10:45.481761 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-smtqw_1437d818-abf4-4602-8a28-e88a76e482a9/cert-manager-controller/0.log" Feb 20 09:10:45 crc kubenswrapper[4948]: I0220 09:10:45.654120 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-9llfv_bb56399a-a840-49af-972f-4f32c91efed7/cert-manager-cainjector/0.log" Feb 20 09:10:45 crc kubenswrapper[4948]: I0220 09:10:45.711084 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-f4bjc_305a17f4-aef3-4036-8fce-3756ff5bbd2f/cert-manager-webhook/0.log" Feb 20 09:10:58 crc kubenswrapper[4948]: I0220 09:10:58.488917 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5c78fc5d65-9mv9z_47461ed8-c51b-4014-bcf0-4d95a0278f85/nmstate-console-plugin/0.log" Feb 20 09:10:58 crc kubenswrapper[4948]: I0220 09:10:58.608757 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-dxdh6_3a869b3a-b757-49ba-9096-2e562d980aae/nmstate-handler/0.log" Feb 20 09:10:58 crc kubenswrapper[4948]: I0220 09:10:58.666656 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-tdst7_a7b34c06-d0c5-4f20-9521-eec500ea2510/nmstate-metrics/0.log" Feb 20 09:10:58 crc kubenswrapper[4948]: I0220 09:10:58.692412 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-tdst7_a7b34c06-d0c5-4f20-9521-eec500ea2510/kube-rbac-proxy/0.log" Feb 20 09:10:58 crc kubenswrapper[4948]: I0220 09:10:58.816068 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-694c9596b7-xqtvg_59277281-0112-45da-a64a-1b1d02b90473/nmstate-operator/0.log" Feb 20 09:10:58 crc kubenswrapper[4948]: I0220 09:10:58.861625 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-866bcb46dc-77sdv_da0e4284-f20d-44a8-b529-e388f15822d3/nmstate-webhook/0.log" Feb 20 09:11:27 crc kubenswrapper[4948]: I0220 09:11:27.705865 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-c6sqk_68ee0185-13d4-45c7-9a39-bb474e9314d0/kube-rbac-proxy/0.log" Feb 20 09:11:27 crc kubenswrapper[4948]: I0220 09:11:27.783424 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-c6sqk_68ee0185-13d4-45c7-9a39-bb474e9314d0/controller/0.log" Feb 20 09:11:27 crc kubenswrapper[4948]: I0220 09:11:27.959565 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-frr-files/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.134463 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-frr-files/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.164319 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-metrics/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.169796 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-reloader/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.215301 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-reloader/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.399921 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-frr-files/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.408036 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-metrics/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.422811 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-reloader/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.440364 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-metrics/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.605323 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-frr-files/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.633253 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-reloader/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.658694 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/controller/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.660088 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/cp-metrics/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.837670 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/kube-rbac-proxy-frr/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.857443 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/frr-metrics/0.log" Feb 20 09:11:28 crc kubenswrapper[4948]: I0220 09:11:28.887725 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/kube-rbac-proxy/0.log" Feb 20 09:11:29 crc kubenswrapper[4948]: I0220 09:11:29.016472 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/reloader/0.log" Feb 20 09:11:29 crc kubenswrapper[4948]: I0220 09:11:29.132898 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-7zsxs_2506d4e2-b434-4fe4-970e-7cd14601677d/frr-k8s-webhook-server/0.log" Feb 20 09:11:29 crc kubenswrapper[4948]: I0220 09:11:29.371889 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-8478c6b8cc-v98zs_b4f86a2e-15f9-441c-953b-49f331d4122e/manager/0.log" Feb 20 09:11:29 crc kubenswrapper[4948]: I0220 09:11:29.440844 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5fcd7d546c-kgpxs_07022b13-7dc1-45d1-a626-91792bd3aa90/webhook-server/0.log" Feb 20 09:11:29 crc kubenswrapper[4948]: I0220 09:11:29.639730 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-h7xhj_2c199acd-2fd3-4b9f-b50c-a266191c4777/kube-rbac-proxy/0.log" Feb 20 09:11:30 crc kubenswrapper[4948]: I0220 09:11:30.139988 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-h7xhj_2c199acd-2fd3-4b9f-b50c-a266191c4777/speaker/0.log" Feb 20 09:11:30 crc kubenswrapper[4948]: I0220 09:11:30.246447 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k66nh_0131d05b-619d-43d8-aa38-ad6e9ce52e11/frr/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.207899 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/util/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.382877 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/util/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.435636 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/pull/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.474397 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/pull/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.570367 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/util/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.598168 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/pull/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.635788 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213r7bct_9e5e2f81-8ea9-4ddb-b2ab-584c40b15916/extract/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.778226 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-utilities/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.971657 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-utilities/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.971682 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-content/0.log" Feb 20 09:11:43 crc kubenswrapper[4948]: I0220 09:11:43.999182 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-content/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.158733 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-content/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.174505 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/extract-utilities/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.351847 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-utilities/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.635595 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-utilities/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.638998 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-content/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.729730 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-content/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.756828 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt6t6_94e3bb0b-818a-491e-8e1e-a1664f3a9f12/registry-server/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.869208 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-utilities/0.log" Feb 20 09:11:44 crc kubenswrapper[4948]: I0220 09:11:44.911955 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/extract-content/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.074101 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/util/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.334627 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/pull/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.344033 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/util/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.402411 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl7nj_0e759dc6-36e8-4a97-b994-e657f1b5d540/registry-server/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.408240 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/pull/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.507856 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/util/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.551238 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/pull/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.584479 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecatqc8w_bb3ec19c-541d-4b32-9f5c-ac61f28bdc2a/extract/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.744738 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-utilities/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.777277 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-tbqzs_1e3d2cd4-4311-4454-b257-20a3caf243b3/marketplace-operator/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.944292 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-content/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.974712 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-content/0.log" Feb 20 09:11:45 crc kubenswrapper[4948]: I0220 09:11:45.981486 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-utilities/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.123896 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-utilities/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.215445 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/extract-content/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.283814 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x7bcm_6ec98112-e189-4a27-a6dd-8cb035596810/registry-server/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.332439 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-utilities/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.447260 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-content/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.484618 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-content/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.504658 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-utilities/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.670905 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-utilities/0.log" Feb 20 09:11:46 crc kubenswrapper[4948]: I0220 09:11:46.681053 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/extract-content/0.log" Feb 20 09:11:47 crc kubenswrapper[4948]: I0220 09:11:47.130722 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbtm5_5ee242c9-9783-44cb-a4dc-1ea50a7ae185/registry-server/0.log" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.034608 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjmh"] Feb 20 09:11:48 crc kubenswrapper[4948]: E0220 09:11:48.035155 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfde212-7150-4b57-b3dd-41715c736426" containerName="container-00" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.035176 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfde212-7150-4b57-b3dd-41715c736426" containerName="container-00" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.035502 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfde212-7150-4b57-b3dd-41715c736426" containerName="container-00" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.037238 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.063773 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjmh"] Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.113642 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4v6l\" (UniqueName: \"kubernetes.io/projected/654b30bf-3b65-4501-9a0e-9ebc33812c73-kube-api-access-g4v6l\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.113734 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-utilities\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.113900 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-catalog-content\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.215375 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-catalog-content\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.215621 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4v6l\" (UniqueName: \"kubernetes.io/projected/654b30bf-3b65-4501-9a0e-9ebc33812c73-kube-api-access-g4v6l\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.215655 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-utilities\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.216077 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-catalog-content\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.216153 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-utilities\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.234935 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4v6l\" (UniqueName: \"kubernetes.io/projected/654b30bf-3b65-4501-9a0e-9ebc33812c73-kube-api-access-g4v6l\") pod \"redhat-marketplace-7sjmh\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.365315 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:48 crc kubenswrapper[4948]: I0220 09:11:48.833940 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjmh"] Feb 20 09:11:49 crc kubenswrapper[4948]: I0220 09:11:49.618245 4948 generic.go:334] "Generic (PLEG): container finished" podID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerID="0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915" exitCode=0 Feb 20 09:11:49 crc kubenswrapper[4948]: I0220 09:11:49.618304 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjmh" event={"ID":"654b30bf-3b65-4501-9a0e-9ebc33812c73","Type":"ContainerDied","Data":"0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915"} Feb 20 09:11:49 crc kubenswrapper[4948]: I0220 09:11:49.618335 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjmh" event={"ID":"654b30bf-3b65-4501-9a0e-9ebc33812c73","Type":"ContainerStarted","Data":"935397e9a001171af02db7a3be1a28e8bc26a789185e6481f73a103b9280bae0"} Feb 20 09:11:49 crc kubenswrapper[4948]: I0220 09:11:49.621397 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 09:11:50 crc kubenswrapper[4948]: I0220 09:11:50.632684 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjmh" event={"ID":"654b30bf-3b65-4501-9a0e-9ebc33812c73","Type":"ContainerStarted","Data":"f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973"} Feb 20 09:11:51 crc kubenswrapper[4948]: I0220 09:11:51.644766 4948 generic.go:334] "Generic (PLEG): container finished" podID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerID="f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973" exitCode=0 Feb 20 09:11:51 crc kubenswrapper[4948]: I0220 09:11:51.644830 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjmh" event={"ID":"654b30bf-3b65-4501-9a0e-9ebc33812c73","Type":"ContainerDied","Data":"f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973"} Feb 20 09:11:52 crc kubenswrapper[4948]: I0220 09:11:52.655645 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjmh" event={"ID":"654b30bf-3b65-4501-9a0e-9ebc33812c73","Type":"ContainerStarted","Data":"e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7"} Feb 20 09:11:52 crc kubenswrapper[4948]: I0220 09:11:52.683021 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7sjmh" podStartSLOduration=2.273941058 podStartE2EDuration="4.682959414s" podCreationTimestamp="2026-02-20 09:11:48 +0000 UTC" firstStartedPulling="2026-02-20 09:11:49.621121087 +0000 UTC m=+3958.595615907" lastFinishedPulling="2026-02-20 09:11:52.030139403 +0000 UTC m=+3961.004634263" observedRunningTime="2026-02-20 09:11:52.674009473 +0000 UTC m=+3961.648504313" watchObservedRunningTime="2026-02-20 09:11:52.682959414 +0000 UTC m=+3961.657454234" Feb 20 09:11:58 crc kubenswrapper[4948]: I0220 09:11:58.366103 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:58 crc kubenswrapper[4948]: I0220 09:11:58.366682 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:58 crc kubenswrapper[4948]: I0220 09:11:58.406233 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:58 crc kubenswrapper[4948]: I0220 09:11:58.757965 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:11:58 crc kubenswrapper[4948]: I0220 09:11:58.817171 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjmh"] Feb 20 09:12:00 crc kubenswrapper[4948]: I0220 09:12:00.725097 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7sjmh" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerName="registry-server" containerID="cri-o://e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7" gracePeriod=2 Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.297466 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.420015 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4v6l\" (UniqueName: \"kubernetes.io/projected/654b30bf-3b65-4501-9a0e-9ebc33812c73-kube-api-access-g4v6l\") pod \"654b30bf-3b65-4501-9a0e-9ebc33812c73\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.420121 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-catalog-content\") pod \"654b30bf-3b65-4501-9a0e-9ebc33812c73\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.420228 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-utilities\") pod \"654b30bf-3b65-4501-9a0e-9ebc33812c73\" (UID: \"654b30bf-3b65-4501-9a0e-9ebc33812c73\") " Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.421602 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-utilities" (OuterVolumeSpecName: "utilities") pod "654b30bf-3b65-4501-9a0e-9ebc33812c73" (UID: "654b30bf-3b65-4501-9a0e-9ebc33812c73"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.426466 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/654b30bf-3b65-4501-9a0e-9ebc33812c73-kube-api-access-g4v6l" (OuterVolumeSpecName: "kube-api-access-g4v6l") pod "654b30bf-3b65-4501-9a0e-9ebc33812c73" (UID: "654b30bf-3b65-4501-9a0e-9ebc33812c73"). InnerVolumeSpecName "kube-api-access-g4v6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.448434 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "654b30bf-3b65-4501-9a0e-9ebc33812c73" (UID: "654b30bf-3b65-4501-9a0e-9ebc33812c73"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.522805 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.523087 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/654b30bf-3b65-4501-9a0e-9ebc33812c73-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.523150 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4v6l\" (UniqueName: \"kubernetes.io/projected/654b30bf-3b65-4501-9a0e-9ebc33812c73-kube-api-access-g4v6l\") on node \"crc\" DevicePath \"\"" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.739926 4948 generic.go:334] "Generic (PLEG): container finished" podID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerID="e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7" exitCode=0 Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.740028 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjmh" event={"ID":"654b30bf-3b65-4501-9a0e-9ebc33812c73","Type":"ContainerDied","Data":"e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7"} Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.740091 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sjmh" event={"ID":"654b30bf-3b65-4501-9a0e-9ebc33812c73","Type":"ContainerDied","Data":"935397e9a001171af02db7a3be1a28e8bc26a789185e6481f73a103b9280bae0"} Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.740131 4948 scope.go:117] "RemoveContainer" containerID="e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.740143 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sjmh" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.760615 4948 scope.go:117] "RemoveContainer" containerID="f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.788160 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjmh"] Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.790367 4948 scope.go:117] "RemoveContainer" containerID="0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.795329 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sjmh"] Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.822881 4948 scope.go:117] "RemoveContainer" containerID="e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7" Feb 20 09:12:01 crc kubenswrapper[4948]: E0220 09:12:01.823343 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7\": container with ID starting with e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7 not found: ID does not exist" containerID="e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.823378 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7"} err="failed to get container status \"e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7\": rpc error: code = NotFound desc = could not find container \"e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7\": container with ID starting with e00180d9024f5ed42ba57eca358c05e3488371a087ae601107ee28e8c4a0c9d7 not found: ID does not exist" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.823398 4948 scope.go:117] "RemoveContainer" containerID="f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973" Feb 20 09:12:01 crc kubenswrapper[4948]: E0220 09:12:01.823681 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973\": container with ID starting with f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973 not found: ID does not exist" containerID="f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.823711 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973"} err="failed to get container status \"f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973\": rpc error: code = NotFound desc = could not find container \"f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973\": container with ID starting with f78e60617dfa8cf4a04999e07015aa7e834279c7e0b99a1ff6c2ec26cb91d973 not found: ID does not exist" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.823732 4948 scope.go:117] "RemoveContainer" containerID="0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915" Feb 20 09:12:01 crc kubenswrapper[4948]: E0220 09:12:01.824013 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915\": container with ID starting with 0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915 not found: ID does not exist" containerID="0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915" Feb 20 09:12:01 crc kubenswrapper[4948]: I0220 09:12:01.824038 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915"} err="failed to get container status \"0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915\": rpc error: code = NotFound desc = could not find container \"0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915\": container with ID starting with 0498136ca7009d716fd78a1af98d3ae2091af6957f0aa8b1253de62c66c87915 not found: ID does not exist" Feb 20 09:12:03 crc kubenswrapper[4948]: I0220 09:12:03.732082 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" path="/var/lib/kubelet/pods/654b30bf-3b65-4501-9a0e-9ebc33812c73/volumes" Feb 20 09:12:16 crc kubenswrapper[4948]: E0220 09:12:16.599168 4948 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.113:40572->38.102.83.113:38489: write tcp 38.102.83.113:40572->38.102.83.113:38489: write: connection reset by peer Feb 20 09:12:38 crc kubenswrapper[4948]: I0220 09:12:38.025188 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:12:38 crc kubenswrapper[4948]: I0220 09:12:38.025698 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:13:08 crc kubenswrapper[4948]: I0220 09:13:08.024993 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:13:08 crc kubenswrapper[4948]: I0220 09:13:08.025740 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:13:31 crc kubenswrapper[4948]: I0220 09:13:31.964009 4948 generic.go:334] "Generic (PLEG): container finished" podID="d81b349b-2494-4abc-b349-44fecd284709" containerID="0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff" exitCode=0 Feb 20 09:13:31 crc kubenswrapper[4948]: I0220 09:13:31.964089 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" event={"ID":"d81b349b-2494-4abc-b349-44fecd284709","Type":"ContainerDied","Data":"0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff"} Feb 20 09:13:31 crc kubenswrapper[4948]: I0220 09:13:31.965381 4948 scope.go:117] "RemoveContainer" containerID="0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff" Feb 20 09:13:32 crc kubenswrapper[4948]: I0220 09:13:32.973328 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ql8t9_must-gather-hvzrp_d81b349b-2494-4abc-b349-44fecd284709/gather/0.log" Feb 20 09:13:38 crc kubenswrapper[4948]: I0220 09:13:38.025180 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:13:38 crc kubenswrapper[4948]: I0220 09:13:38.025821 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:13:38 crc kubenswrapper[4948]: I0220 09:13:38.025919 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 09:13:38 crc kubenswrapper[4948]: I0220 09:13:38.028879 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5bbfc48505968074b16a56aea08c035f5c0687419ea24b66d69a44ebd2ab9d7d"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 09:13:38 crc kubenswrapper[4948]: I0220 09:13:38.028996 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://5bbfc48505968074b16a56aea08c035f5c0687419ea24b66d69a44ebd2ab9d7d" gracePeriod=600 Feb 20 09:13:39 crc kubenswrapper[4948]: I0220 09:13:39.034930 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="5bbfc48505968074b16a56aea08c035f5c0687419ea24b66d69a44ebd2ab9d7d" exitCode=0 Feb 20 09:13:39 crc kubenswrapper[4948]: I0220 09:13:39.034989 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"5bbfc48505968074b16a56aea08c035f5c0687419ea24b66d69a44ebd2ab9d7d"} Feb 20 09:13:39 crc kubenswrapper[4948]: I0220 09:13:39.035498 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerStarted","Data":"2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5"} Feb 20 09:13:39 crc kubenswrapper[4948]: I0220 09:13:39.035523 4948 scope.go:117] "RemoveContainer" containerID="6da9ae9ed55d6ad4892f71f9c96b347ef289ef82b2ec2dd6ac6a34dc2f33a4d2" Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.039438 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ql8t9/must-gather-hvzrp"] Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.040556 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" podUID="d81b349b-2494-4abc-b349-44fecd284709" containerName="copy" containerID="cri-o://d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf" gracePeriod=2 Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.047125 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ql8t9/must-gather-hvzrp"] Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.489147 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ql8t9_must-gather-hvzrp_d81b349b-2494-4abc-b349-44fecd284709/copy/0.log" Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.490107 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.602837 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d81b349b-2494-4abc-b349-44fecd284709-must-gather-output\") pod \"d81b349b-2494-4abc-b349-44fecd284709\" (UID: \"d81b349b-2494-4abc-b349-44fecd284709\") " Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.603091 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2bcm\" (UniqueName: \"kubernetes.io/projected/d81b349b-2494-4abc-b349-44fecd284709-kube-api-access-v2bcm\") pod \"d81b349b-2494-4abc-b349-44fecd284709\" (UID: \"d81b349b-2494-4abc-b349-44fecd284709\") " Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.610136 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d81b349b-2494-4abc-b349-44fecd284709-kube-api-access-v2bcm" (OuterVolumeSpecName: "kube-api-access-v2bcm") pod "d81b349b-2494-4abc-b349-44fecd284709" (UID: "d81b349b-2494-4abc-b349-44fecd284709"). InnerVolumeSpecName "kube-api-access-v2bcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.704765 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2bcm\" (UniqueName: \"kubernetes.io/projected/d81b349b-2494-4abc-b349-44fecd284709-kube-api-access-v2bcm\") on node \"crc\" DevicePath \"\"" Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.770661 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d81b349b-2494-4abc-b349-44fecd284709-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d81b349b-2494-4abc-b349-44fecd284709" (UID: "d81b349b-2494-4abc-b349-44fecd284709"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:13:45 crc kubenswrapper[4948]: I0220 09:13:45.806896 4948 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d81b349b-2494-4abc-b349-44fecd284709-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.166722 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ql8t9_must-gather-hvzrp_d81b349b-2494-4abc-b349-44fecd284709/copy/0.log" Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.167513 4948 generic.go:334] "Generic (PLEG): container finished" podID="d81b349b-2494-4abc-b349-44fecd284709" containerID="d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf" exitCode=143 Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.167585 4948 scope.go:117] "RemoveContainer" containerID="d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf" Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.167725 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ql8t9/must-gather-hvzrp" Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.191045 4948 scope.go:117] "RemoveContainer" containerID="0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff" Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.287325 4948 scope.go:117] "RemoveContainer" containerID="d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf" Feb 20 09:13:46 crc kubenswrapper[4948]: E0220 09:13:46.287795 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf\": container with ID starting with d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf not found: ID does not exist" containerID="d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf" Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.287823 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf"} err="failed to get container status \"d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf\": rpc error: code = NotFound desc = could not find container \"d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf\": container with ID starting with d49cb379a847549834cd028d6c373359633cf1e9d6c301526cf59089fd2bfebf not found: ID does not exist" Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.287848 4948 scope.go:117] "RemoveContainer" containerID="0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff" Feb 20 09:13:46 crc kubenswrapper[4948]: E0220 09:13:46.288123 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff\": container with ID starting with 0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff not found: ID does not exist" containerID="0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff" Feb 20 09:13:46 crc kubenswrapper[4948]: I0220 09:13:46.288147 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff"} err="failed to get container status \"0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff\": rpc error: code = NotFound desc = could not find container \"0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff\": container with ID starting with 0b5814bb586cbca972e209566bf544539a27cba45472fac01d96fe47fe540cff not found: ID does not exist" Feb 20 09:13:47 crc kubenswrapper[4948]: I0220 09:13:47.745410 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d81b349b-2494-4abc-b349-44fecd284709" path="/var/lib/kubelet/pods/d81b349b-2494-4abc-b349-44fecd284709/volumes" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.207739 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c"] Feb 20 09:15:00 crc kubenswrapper[4948]: E0220 09:15:00.208803 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerName="extract-content" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.208821 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerName="extract-content" Feb 20 09:15:00 crc kubenswrapper[4948]: E0220 09:15:00.208834 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d81b349b-2494-4abc-b349-44fecd284709" containerName="gather" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.208843 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d81b349b-2494-4abc-b349-44fecd284709" containerName="gather" Feb 20 09:15:00 crc kubenswrapper[4948]: E0220 09:15:00.208875 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerName="registry-server" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.208889 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerName="registry-server" Feb 20 09:15:00 crc kubenswrapper[4948]: E0220 09:15:00.208905 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d81b349b-2494-4abc-b349-44fecd284709" containerName="copy" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.208914 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d81b349b-2494-4abc-b349-44fecd284709" containerName="copy" Feb 20 09:15:00 crc kubenswrapper[4948]: E0220 09:15:00.208946 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerName="extract-utilities" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.208958 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerName="extract-utilities" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.209220 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="654b30bf-3b65-4501-9a0e-9ebc33812c73" containerName="registry-server" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.209250 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d81b349b-2494-4abc-b349-44fecd284709" containerName="copy" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.209276 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d81b349b-2494-4abc-b349-44fecd284709" containerName="gather" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.210150 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.212657 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.212778 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.231541 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c"] Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.358952 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb9ml\" (UniqueName: \"kubernetes.io/projected/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-kube-api-access-wb9ml\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.359213 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-secret-volume\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.359589 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-config-volume\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.461541 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb9ml\" (UniqueName: \"kubernetes.io/projected/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-kube-api-access-wb9ml\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.462010 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-secret-volume\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.462257 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-config-volume\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.463110 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-config-volume\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.470541 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-secret-volume\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.479157 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb9ml\" (UniqueName: \"kubernetes.io/projected/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-kube-api-access-wb9ml\") pod \"collect-profiles-29526315-h789c\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:00 crc kubenswrapper[4948]: I0220 09:15:00.537768 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:01 crc kubenswrapper[4948]: I0220 09:15:01.033829 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c"] Feb 20 09:15:01 crc kubenswrapper[4948]: I0220 09:15:01.937212 4948 generic.go:334] "Generic (PLEG): container finished" podID="6b98fa20-139d-4a56-b4e8-d32209bf4fb4" containerID="a70d23d0cc71f3b65f68123bbc28e9d24e5f9ab9fb49d3e0ece37db0403b1cdc" exitCode=0 Feb 20 09:15:01 crc kubenswrapper[4948]: I0220 09:15:01.937268 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" event={"ID":"6b98fa20-139d-4a56-b4e8-d32209bf4fb4","Type":"ContainerDied","Data":"a70d23d0cc71f3b65f68123bbc28e9d24e5f9ab9fb49d3e0ece37db0403b1cdc"} Feb 20 09:15:01 crc kubenswrapper[4948]: I0220 09:15:01.937501 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" event={"ID":"6b98fa20-139d-4a56-b4e8-d32209bf4fb4","Type":"ContainerStarted","Data":"03e4e4d97c682186adfd9949701d3c54e01060de77e731ddb5f428f35ac442a8"} Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.328012 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.421948 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-config-volume\") pod \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.422169 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb9ml\" (UniqueName: \"kubernetes.io/projected/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-kube-api-access-wb9ml\") pod \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.422206 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-secret-volume\") pod \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\" (UID: \"6b98fa20-139d-4a56-b4e8-d32209bf4fb4\") " Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.422808 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-config-volume" (OuterVolumeSpecName: "config-volume") pod "6b98fa20-139d-4a56-b4e8-d32209bf4fb4" (UID: "6b98fa20-139d-4a56-b4e8-d32209bf4fb4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.429047 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6b98fa20-139d-4a56-b4e8-d32209bf4fb4" (UID: "6b98fa20-139d-4a56-b4e8-d32209bf4fb4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.429284 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-kube-api-access-wb9ml" (OuterVolumeSpecName: "kube-api-access-wb9ml") pod "6b98fa20-139d-4a56-b4e8-d32209bf4fb4" (UID: "6b98fa20-139d-4a56-b4e8-d32209bf4fb4"). InnerVolumeSpecName "kube-api-access-wb9ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.524723 4948 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.524997 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb9ml\" (UniqueName: \"kubernetes.io/projected/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-kube-api-access-wb9ml\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.525008 4948 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b98fa20-139d-4a56-b4e8-d32209bf4fb4-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.971661 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" event={"ID":"6b98fa20-139d-4a56-b4e8-d32209bf4fb4","Type":"ContainerDied","Data":"03e4e4d97c682186adfd9949701d3c54e01060de77e731ddb5f428f35ac442a8"} Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.971720 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03e4e4d97c682186adfd9949701d3c54e01060de77e731ddb5f428f35ac442a8" Feb 20 09:15:03 crc kubenswrapper[4948]: I0220 09:15:03.971743 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-h789c" Feb 20 09:15:04 crc kubenswrapper[4948]: I0220 09:15:04.421054 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm"] Feb 20 09:15:04 crc kubenswrapper[4948]: I0220 09:15:04.431928 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526270-98wxm"] Feb 20 09:15:05 crc kubenswrapper[4948]: I0220 09:15:05.738487 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aaa3d0b-4371-404a-8d94-9a7158fff416" path="/var/lib/kubelet/pods/2aaa3d0b-4371-404a-8d94-9a7158fff416/volumes" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.377131 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9mdkj"] Feb 20 09:15:15 crc kubenswrapper[4948]: E0220 09:15:15.378644 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b98fa20-139d-4a56-b4e8-d32209bf4fb4" containerName="collect-profiles" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.378671 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b98fa20-139d-4a56-b4e8-d32209bf4fb4" containerName="collect-profiles" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.379221 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b98fa20-139d-4a56-b4e8-d32209bf4fb4" containerName="collect-profiles" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.382246 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.404020 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9mdkj"] Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.499132 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-catalog-content\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.499278 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-utilities\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.499603 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwrms\" (UniqueName: \"kubernetes.io/projected/1443dd18-dcab-42e8-8592-63e1ae18a6ef-kube-api-access-xwrms\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.601802 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-catalog-content\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.601883 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-utilities\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.602027 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwrms\" (UniqueName: \"kubernetes.io/projected/1443dd18-dcab-42e8-8592-63e1ae18a6ef-kube-api-access-xwrms\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.602552 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-catalog-content\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.602732 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-utilities\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.626694 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwrms\" (UniqueName: \"kubernetes.io/projected/1443dd18-dcab-42e8-8592-63e1ae18a6ef-kube-api-access-xwrms\") pod \"certified-operators-9mdkj\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:15 crc kubenswrapper[4948]: I0220 09:15:15.711210 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:16 crc kubenswrapper[4948]: I0220 09:15:16.222299 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9mdkj"] Feb 20 09:15:17 crc kubenswrapper[4948]: I0220 09:15:17.142434 4948 generic.go:334] "Generic (PLEG): container finished" podID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerID="6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c" exitCode=0 Feb 20 09:15:17 crc kubenswrapper[4948]: I0220 09:15:17.142500 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mdkj" event={"ID":"1443dd18-dcab-42e8-8592-63e1ae18a6ef","Type":"ContainerDied","Data":"6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c"} Feb 20 09:15:17 crc kubenswrapper[4948]: I0220 09:15:17.142557 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mdkj" event={"ID":"1443dd18-dcab-42e8-8592-63e1ae18a6ef","Type":"ContainerStarted","Data":"f5de5732c2b7595df1124e25393432adf4d492acc7ff2a3bfd2cf5030afc2d33"} Feb 20 09:15:18 crc kubenswrapper[4948]: I0220 09:15:18.158143 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mdkj" event={"ID":"1443dd18-dcab-42e8-8592-63e1ae18a6ef","Type":"ContainerStarted","Data":"e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042"} Feb 20 09:15:19 crc kubenswrapper[4948]: I0220 09:15:19.171320 4948 generic.go:334] "Generic (PLEG): container finished" podID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerID="e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042" exitCode=0 Feb 20 09:15:19 crc kubenswrapper[4948]: I0220 09:15:19.171402 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mdkj" event={"ID":"1443dd18-dcab-42e8-8592-63e1ae18a6ef","Type":"ContainerDied","Data":"e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042"} Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.186127 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mdkj" event={"ID":"1443dd18-dcab-42e8-8592-63e1ae18a6ef","Type":"ContainerStarted","Data":"011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a"} Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.208519 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9mdkj" podStartSLOduration=2.743889042 podStartE2EDuration="5.20849947s" podCreationTimestamp="2026-02-20 09:15:15 +0000 UTC" firstStartedPulling="2026-02-20 09:15:17.145506114 +0000 UTC m=+4166.120000974" lastFinishedPulling="2026-02-20 09:15:19.610116542 +0000 UTC m=+4168.584611402" observedRunningTime="2026-02-20 09:15:20.206228754 +0000 UTC m=+4169.180723614" watchObservedRunningTime="2026-02-20 09:15:20.20849947 +0000 UTC m=+4169.182994290" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.349311 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d28wc"] Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.353300 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.361643 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d28wc"] Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.513305 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-utilities\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.513471 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl724\" (UniqueName: \"kubernetes.io/projected/4425c869-8c06-4583-bd41-9fbb012a893e-kube-api-access-rl724\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.513508 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-catalog-content\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.614840 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl724\" (UniqueName: \"kubernetes.io/projected/4425c869-8c06-4583-bd41-9fbb012a893e-kube-api-access-rl724\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.614920 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-catalog-content\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.615113 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-utilities\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.615599 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-utilities\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.615695 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-catalog-content\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.639710 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl724\" (UniqueName: \"kubernetes.io/projected/4425c869-8c06-4583-bd41-9fbb012a893e-kube-api-access-rl724\") pod \"community-operators-d28wc\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:20 crc kubenswrapper[4948]: I0220 09:15:20.686443 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:21 crc kubenswrapper[4948]: I0220 09:15:21.214002 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d28wc"] Feb 20 09:15:21 crc kubenswrapper[4948]: W0220 09:15:21.222405 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4425c869_8c06_4583_bd41_9fbb012a893e.slice/crio-5b8038ad2561ab00e000949c744573df91dd4c5894d48f3525defea5f9208cb0 WatchSource:0}: Error finding container 5b8038ad2561ab00e000949c744573df91dd4c5894d48f3525defea5f9208cb0: Status 404 returned error can't find the container with id 5b8038ad2561ab00e000949c744573df91dd4c5894d48f3525defea5f9208cb0 Feb 20 09:15:22 crc kubenswrapper[4948]: I0220 09:15:22.212331 4948 generic.go:334] "Generic (PLEG): container finished" podID="4425c869-8c06-4583-bd41-9fbb012a893e" containerID="5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722" exitCode=0 Feb 20 09:15:22 crc kubenswrapper[4948]: I0220 09:15:22.213446 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d28wc" event={"ID":"4425c869-8c06-4583-bd41-9fbb012a893e","Type":"ContainerDied","Data":"5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722"} Feb 20 09:15:22 crc kubenswrapper[4948]: I0220 09:15:22.213684 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d28wc" event={"ID":"4425c869-8c06-4583-bd41-9fbb012a893e","Type":"ContainerStarted","Data":"5b8038ad2561ab00e000949c744573df91dd4c5894d48f3525defea5f9208cb0"} Feb 20 09:15:24 crc kubenswrapper[4948]: I0220 09:15:24.241469 4948 generic.go:334] "Generic (PLEG): container finished" podID="4425c869-8c06-4583-bd41-9fbb012a893e" containerID="3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242" exitCode=0 Feb 20 09:15:24 crc kubenswrapper[4948]: I0220 09:15:24.242018 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d28wc" event={"ID":"4425c869-8c06-4583-bd41-9fbb012a893e","Type":"ContainerDied","Data":"3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242"} Feb 20 09:15:25 crc kubenswrapper[4948]: I0220 09:15:25.257588 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d28wc" event={"ID":"4425c869-8c06-4583-bd41-9fbb012a893e","Type":"ContainerStarted","Data":"edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be"} Feb 20 09:15:25 crc kubenswrapper[4948]: I0220 09:15:25.287798 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d28wc" podStartSLOduration=2.886426821 podStartE2EDuration="5.287775047s" podCreationTimestamp="2026-02-20 09:15:20 +0000 UTC" firstStartedPulling="2026-02-20 09:15:22.224682419 +0000 UTC m=+4171.199177279" lastFinishedPulling="2026-02-20 09:15:24.626030645 +0000 UTC m=+4173.600525505" observedRunningTime="2026-02-20 09:15:25.277046352 +0000 UTC m=+4174.251541182" watchObservedRunningTime="2026-02-20 09:15:25.287775047 +0000 UTC m=+4174.262269887" Feb 20 09:15:25 crc kubenswrapper[4948]: I0220 09:15:25.712294 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:25 crc kubenswrapper[4948]: I0220 09:15:25.712357 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:25 crc kubenswrapper[4948]: I0220 09:15:25.774882 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:26 crc kubenswrapper[4948]: I0220 09:15:26.353486 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:26 crc kubenswrapper[4948]: I0220 09:15:26.930437 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9mdkj"] Feb 20 09:15:28 crc kubenswrapper[4948]: I0220 09:15:28.290290 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9mdkj" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerName="registry-server" containerID="cri-o://011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a" gracePeriod=2 Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.267413 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.301643 4948 generic.go:334] "Generic (PLEG): container finished" podID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerID="011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a" exitCode=0 Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.301690 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mdkj" event={"ID":"1443dd18-dcab-42e8-8592-63e1ae18a6ef","Type":"ContainerDied","Data":"011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a"} Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.301706 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9mdkj" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.301734 4948 scope.go:117] "RemoveContainer" containerID="011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.301722 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9mdkj" event={"ID":"1443dd18-dcab-42e8-8592-63e1ae18a6ef","Type":"ContainerDied","Data":"f5de5732c2b7595df1124e25393432adf4d492acc7ff2a3bfd2cf5030afc2d33"} Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.321212 4948 scope.go:117] "RemoveContainer" containerID="e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.348924 4948 scope.go:117] "RemoveContainer" containerID="6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.386185 4948 scope.go:117] "RemoveContainer" containerID="011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a" Feb 20 09:15:29 crc kubenswrapper[4948]: E0220 09:15:29.386662 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a\": container with ID starting with 011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a not found: ID does not exist" containerID="011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.386703 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a"} err="failed to get container status \"011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a\": rpc error: code = NotFound desc = could not find container \"011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a\": container with ID starting with 011cc2bc06960fc14c7f8fb4e578e2b7d990552c5f3d14d6bd6095826445bf0a not found: ID does not exist" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.386729 4948 scope.go:117] "RemoveContainer" containerID="e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042" Feb 20 09:15:29 crc kubenswrapper[4948]: E0220 09:15:29.387356 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042\": container with ID starting with e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042 not found: ID does not exist" containerID="e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.387455 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042"} err="failed to get container status \"e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042\": rpc error: code = NotFound desc = could not find container \"e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042\": container with ID starting with e4d2e8ef702be4e555465ec7a1aaea879fb47d956b0f098c18df270181ca9042 not found: ID does not exist" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.387511 4948 scope.go:117] "RemoveContainer" containerID="6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c" Feb 20 09:15:29 crc kubenswrapper[4948]: E0220 09:15:29.387877 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c\": container with ID starting with 6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c not found: ID does not exist" containerID="6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.387943 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c"} err="failed to get container status \"6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c\": rpc error: code = NotFound desc = could not find container \"6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c\": container with ID starting with 6cacdffb656dd31edb857180139ac055fde2c217ffc1c87e516b8f4b9092b59c not found: ID does not exist" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.406559 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-utilities\") pod \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.406688 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-catalog-content\") pod \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.406734 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwrms\" (UniqueName: \"kubernetes.io/projected/1443dd18-dcab-42e8-8592-63e1ae18a6ef-kube-api-access-xwrms\") pod \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\" (UID: \"1443dd18-dcab-42e8-8592-63e1ae18a6ef\") " Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.417385 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1443dd18-dcab-42e8-8592-63e1ae18a6ef-kube-api-access-xwrms" (OuterVolumeSpecName: "kube-api-access-xwrms") pod "1443dd18-dcab-42e8-8592-63e1ae18a6ef" (UID: "1443dd18-dcab-42e8-8592-63e1ae18a6ef"). InnerVolumeSpecName "kube-api-access-xwrms". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.418538 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-utilities" (OuterVolumeSpecName: "utilities") pod "1443dd18-dcab-42e8-8592-63e1ae18a6ef" (UID: "1443dd18-dcab-42e8-8592-63e1ae18a6ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.477374 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1443dd18-dcab-42e8-8592-63e1ae18a6ef" (UID: "1443dd18-dcab-42e8-8592-63e1ae18a6ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.508613 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.508639 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1443dd18-dcab-42e8-8592-63e1ae18a6ef-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.508651 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwrms\" (UniqueName: \"kubernetes.io/projected/1443dd18-dcab-42e8-8592-63e1ae18a6ef-kube-api-access-xwrms\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.671176 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9mdkj"] Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.681744 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9mdkj"] Feb 20 09:15:29 crc kubenswrapper[4948]: I0220 09:15:29.746363 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" path="/var/lib/kubelet/pods/1443dd18-dcab-42e8-8592-63e1ae18a6ef/volumes" Feb 20 09:15:30 crc kubenswrapper[4948]: I0220 09:15:30.686546 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:30 crc kubenswrapper[4948]: I0220 09:15:30.686938 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:30 crc kubenswrapper[4948]: I0220 09:15:30.756412 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:31 crc kubenswrapper[4948]: I0220 09:15:31.399025 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:32 crc kubenswrapper[4948]: I0220 09:15:32.534835 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d28wc"] Feb 20 09:15:33 crc kubenswrapper[4948]: I0220 09:15:33.351282 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d28wc" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" containerName="registry-server" containerID="cri-o://edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be" gracePeriod=2 Feb 20 09:15:33 crc kubenswrapper[4948]: I0220 09:15:33.866063 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.007469 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rl724\" (UniqueName: \"kubernetes.io/projected/4425c869-8c06-4583-bd41-9fbb012a893e-kube-api-access-rl724\") pod \"4425c869-8c06-4583-bd41-9fbb012a893e\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.007683 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-catalog-content\") pod \"4425c869-8c06-4583-bd41-9fbb012a893e\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.007845 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-utilities\") pod \"4425c869-8c06-4583-bd41-9fbb012a893e\" (UID: \"4425c869-8c06-4583-bd41-9fbb012a893e\") " Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.009123 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-utilities" (OuterVolumeSpecName: "utilities") pod "4425c869-8c06-4583-bd41-9fbb012a893e" (UID: "4425c869-8c06-4583-bd41-9fbb012a893e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.013865 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.016354 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4425c869-8c06-4583-bd41-9fbb012a893e-kube-api-access-rl724" (OuterVolumeSpecName: "kube-api-access-rl724") pod "4425c869-8c06-4583-bd41-9fbb012a893e" (UID: "4425c869-8c06-4583-bd41-9fbb012a893e"). InnerVolumeSpecName "kube-api-access-rl724". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.117503 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rl724\" (UniqueName: \"kubernetes.io/projected/4425c869-8c06-4583-bd41-9fbb012a893e-kube-api-access-rl724\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.266878 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4425c869-8c06-4583-bd41-9fbb012a893e" (UID: "4425c869-8c06-4583-bd41-9fbb012a893e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.322046 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4425c869-8c06-4583-bd41-9fbb012a893e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.368300 4948 generic.go:334] "Generic (PLEG): container finished" podID="4425c869-8c06-4583-bd41-9fbb012a893e" containerID="edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be" exitCode=0 Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.368347 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d28wc" event={"ID":"4425c869-8c06-4583-bd41-9fbb012a893e","Type":"ContainerDied","Data":"edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be"} Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.368408 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d28wc" event={"ID":"4425c869-8c06-4583-bd41-9fbb012a893e","Type":"ContainerDied","Data":"5b8038ad2561ab00e000949c744573df91dd4c5894d48f3525defea5f9208cb0"} Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.368434 4948 scope.go:117] "RemoveContainer" containerID="edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.368540 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d28wc" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.405672 4948 scope.go:117] "RemoveContainer" containerID="3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.425834 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d28wc"] Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.439149 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d28wc"] Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.445619 4948 scope.go:117] "RemoveContainer" containerID="5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.490791 4948 scope.go:117] "RemoveContainer" containerID="edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be" Feb 20 09:15:34 crc kubenswrapper[4948]: E0220 09:15:34.491414 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be\": container with ID starting with edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be not found: ID does not exist" containerID="edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.491662 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be"} err="failed to get container status \"edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be\": rpc error: code = NotFound desc = could not find container \"edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be\": container with ID starting with edffd0bfb746bfdf03464481477fac6fcf1088c83d307d3e9e259c8480f089be not found: ID does not exist" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.491891 4948 scope.go:117] "RemoveContainer" containerID="3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242" Feb 20 09:15:34 crc kubenswrapper[4948]: E0220 09:15:34.493308 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242\": container with ID starting with 3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242 not found: ID does not exist" containerID="3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.493350 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242"} err="failed to get container status \"3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242\": rpc error: code = NotFound desc = could not find container \"3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242\": container with ID starting with 3cc32df17d4cb5705d3995090fdf4ac05ae48ad237b349fcecac92279bfda242 not found: ID does not exist" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.493399 4948 scope.go:117] "RemoveContainer" containerID="5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722" Feb 20 09:15:34 crc kubenswrapper[4948]: E0220 09:15:34.493784 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722\": container with ID starting with 5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722 not found: ID does not exist" containerID="5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722" Feb 20 09:15:34 crc kubenswrapper[4948]: I0220 09:15:34.493821 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722"} err="failed to get container status \"5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722\": rpc error: code = NotFound desc = could not find container \"5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722\": container with ID starting with 5ddb552b30bf71c7600bd2cc910fb0b0c766c877286178ccb42d4a02b77e9722 not found: ID does not exist" Feb 20 09:15:35 crc kubenswrapper[4948]: I0220 09:15:35.736680 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" path="/var/lib/kubelet/pods/4425c869-8c06-4583-bd41-9fbb012a893e/volumes" Feb 20 09:15:38 crc kubenswrapper[4948]: I0220 09:15:38.025477 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:15:38 crc kubenswrapper[4948]: I0220 09:15:38.026316 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:16:03 crc kubenswrapper[4948]: I0220 09:16:03.417886 4948 scope.go:117] "RemoveContainer" containerID="21765550ece8828e59342442ce6d5be7d08bd1dbcd30f8abbcc286c1bfd5f713" Feb 20 09:16:08 crc kubenswrapper[4948]: I0220 09:16:08.025256 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:16:08 crc kubenswrapper[4948]: I0220 09:16:08.026333 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.992121 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hrvz7"] Feb 20 09:16:20 crc kubenswrapper[4948]: E0220 09:16:20.993332 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerName="registry-server" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.993350 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerName="registry-server" Feb 20 09:16:20 crc kubenswrapper[4948]: E0220 09:16:20.993369 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" containerName="registry-server" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.993378 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" containerName="registry-server" Feb 20 09:16:20 crc kubenswrapper[4948]: E0220 09:16:20.993398 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerName="extract-content" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.993408 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerName="extract-content" Feb 20 09:16:20 crc kubenswrapper[4948]: E0220 09:16:20.993431 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerName="extract-utilities" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.993439 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerName="extract-utilities" Feb 20 09:16:20 crc kubenswrapper[4948]: E0220 09:16:20.993458 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" containerName="extract-content" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.993466 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" containerName="extract-content" Feb 20 09:16:20 crc kubenswrapper[4948]: E0220 09:16:20.993485 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" containerName="extract-utilities" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.993494 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" containerName="extract-utilities" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.993721 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="4425c869-8c06-4583-bd41-9fbb012a893e" containerName="registry-server" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.993738 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="1443dd18-dcab-42e8-8592-63e1ae18a6ef" containerName="registry-server" Feb 20 09:16:20 crc kubenswrapper[4948]: I0220 09:16:20.998653 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.016353 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hrvz7"] Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.160239 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7cqf\" (UniqueName: \"kubernetes.io/projected/69047018-e836-408d-9ff4-f6d9ee4a37b2-kube-api-access-m7cqf\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.160354 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-catalog-content\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.160489 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-utilities\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.261835 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7cqf\" (UniqueName: \"kubernetes.io/projected/69047018-e836-408d-9ff4-f6d9ee4a37b2-kube-api-access-m7cqf\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.262272 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-catalog-content\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.262444 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-utilities\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.262899 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-catalog-content\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.263054 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-utilities\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.289797 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7cqf\" (UniqueName: \"kubernetes.io/projected/69047018-e836-408d-9ff4-f6d9ee4a37b2-kube-api-access-m7cqf\") pod \"redhat-operators-hrvz7\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:21 crc kubenswrapper[4948]: I0220 09:16:21.347049 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:22 crc kubenswrapper[4948]: I0220 09:16:21.822914 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hrvz7"] Feb 20 09:16:22 crc kubenswrapper[4948]: I0220 09:16:22.914785 4948 generic.go:334] "Generic (PLEG): container finished" podID="69047018-e836-408d-9ff4-f6d9ee4a37b2" containerID="3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438" exitCode=0 Feb 20 09:16:22 crc kubenswrapper[4948]: I0220 09:16:22.914859 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvz7" event={"ID":"69047018-e836-408d-9ff4-f6d9ee4a37b2","Type":"ContainerDied","Data":"3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438"} Feb 20 09:16:22 crc kubenswrapper[4948]: I0220 09:16:22.915251 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvz7" event={"ID":"69047018-e836-408d-9ff4-f6d9ee4a37b2","Type":"ContainerStarted","Data":"f45caaea490893f7fa52af61c5e11466b24ca3554bf9aff6751d648751c80151"} Feb 20 09:16:23 crc kubenswrapper[4948]: I0220 09:16:23.926673 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvz7" event={"ID":"69047018-e836-408d-9ff4-f6d9ee4a37b2","Type":"ContainerStarted","Data":"69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758"} Feb 20 09:16:26 crc kubenswrapper[4948]: I0220 09:16:26.959235 4948 generic.go:334] "Generic (PLEG): container finished" podID="69047018-e836-408d-9ff4-f6d9ee4a37b2" containerID="69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758" exitCode=0 Feb 20 09:16:26 crc kubenswrapper[4948]: I0220 09:16:26.959286 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvz7" event={"ID":"69047018-e836-408d-9ff4-f6d9ee4a37b2","Type":"ContainerDied","Data":"69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758"} Feb 20 09:16:27 crc kubenswrapper[4948]: I0220 09:16:27.970744 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvz7" event={"ID":"69047018-e836-408d-9ff4-f6d9ee4a37b2","Type":"ContainerStarted","Data":"18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32"} Feb 20 09:16:27 crc kubenswrapper[4948]: I0220 09:16:27.996637 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hrvz7" podStartSLOduration=3.569465246 podStartE2EDuration="7.996611869s" podCreationTimestamp="2026-02-20 09:16:20 +0000 UTC" firstStartedPulling="2026-02-20 09:16:22.9168677 +0000 UTC m=+4231.891362530" lastFinishedPulling="2026-02-20 09:16:27.344014303 +0000 UTC m=+4236.318509153" observedRunningTime="2026-02-20 09:16:27.99260372 +0000 UTC m=+4236.967098580" watchObservedRunningTime="2026-02-20 09:16:27.996611869 +0000 UTC m=+4236.971106699" Feb 20 09:16:31 crc kubenswrapper[4948]: I0220 09:16:31.347789 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:31 crc kubenswrapper[4948]: I0220 09:16:31.348521 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:32 crc kubenswrapper[4948]: I0220 09:16:32.430174 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hrvz7" podUID="69047018-e836-408d-9ff4-f6d9ee4a37b2" containerName="registry-server" probeResult="failure" output=< Feb 20 09:16:32 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Feb 20 09:16:32 crc kubenswrapper[4948]: > Feb 20 09:16:38 crc kubenswrapper[4948]: I0220 09:16:38.025306 4948 patch_prober.go:28] interesting pod/machine-config-daemon-h4ww2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:16:38 crc kubenswrapper[4948]: I0220 09:16:38.025926 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:16:38 crc kubenswrapper[4948]: I0220 09:16:38.026007 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" Feb 20 09:16:38 crc kubenswrapper[4948]: I0220 09:16:38.026790 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5"} pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 09:16:38 crc kubenswrapper[4948]: I0220 09:16:38.026851 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerName="machine-config-daemon" containerID="cri-o://2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5" gracePeriod=600 Feb 20 09:16:38 crc kubenswrapper[4948]: E0220 09:16:38.154363 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:16:39 crc kubenswrapper[4948]: I0220 09:16:39.093727 4948 generic.go:334] "Generic (PLEG): container finished" podID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" containerID="2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5" exitCode=0 Feb 20 09:16:39 crc kubenswrapper[4948]: I0220 09:16:39.093769 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" event={"ID":"84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee","Type":"ContainerDied","Data":"2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5"} Feb 20 09:16:39 crc kubenswrapper[4948]: I0220 09:16:39.093800 4948 scope.go:117] "RemoveContainer" containerID="5bbfc48505968074b16a56aea08c035f5c0687419ea24b66d69a44ebd2ab9d7d" Feb 20 09:16:39 crc kubenswrapper[4948]: I0220 09:16:39.094731 4948 scope.go:117] "RemoveContainer" containerID="2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5" Feb 20 09:16:39 crc kubenswrapper[4948]: E0220 09:16:39.095355 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:16:41 crc kubenswrapper[4948]: I0220 09:16:41.415158 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:41 crc kubenswrapper[4948]: I0220 09:16:41.471857 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:41 crc kubenswrapper[4948]: I0220 09:16:41.656562 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hrvz7"] Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.131219 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hrvz7" podUID="69047018-e836-408d-9ff4-f6d9ee4a37b2" containerName="registry-server" containerID="cri-o://18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32" gracePeriod=2 Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.661330 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.755153 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-catalog-content\") pod \"69047018-e836-408d-9ff4-f6d9ee4a37b2\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.755236 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7cqf\" (UniqueName: \"kubernetes.io/projected/69047018-e836-408d-9ff4-f6d9ee4a37b2-kube-api-access-m7cqf\") pod \"69047018-e836-408d-9ff4-f6d9ee4a37b2\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.755520 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-utilities\") pod \"69047018-e836-408d-9ff4-f6d9ee4a37b2\" (UID: \"69047018-e836-408d-9ff4-f6d9ee4a37b2\") " Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.756815 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-utilities" (OuterVolumeSpecName: "utilities") pod "69047018-e836-408d-9ff4-f6d9ee4a37b2" (UID: "69047018-e836-408d-9ff4-f6d9ee4a37b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.765561 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69047018-e836-408d-9ff4-f6d9ee4a37b2-kube-api-access-m7cqf" (OuterVolumeSpecName: "kube-api-access-m7cqf") pod "69047018-e836-408d-9ff4-f6d9ee4a37b2" (UID: "69047018-e836-408d-9ff4-f6d9ee4a37b2"). InnerVolumeSpecName "kube-api-access-m7cqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.858465 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7cqf\" (UniqueName: \"kubernetes.io/projected/69047018-e836-408d-9ff4-f6d9ee4a37b2-kube-api-access-m7cqf\") on node \"crc\" DevicePath \"\"" Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.858525 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.905129 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69047018-e836-408d-9ff4-f6d9ee4a37b2" (UID: "69047018-e836-408d-9ff4-f6d9ee4a37b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:16:43 crc kubenswrapper[4948]: I0220 09:16:43.960049 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69047018-e836-408d-9ff4-f6d9ee4a37b2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.144407 4948 generic.go:334] "Generic (PLEG): container finished" podID="69047018-e836-408d-9ff4-f6d9ee4a37b2" containerID="18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32" exitCode=0 Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.144461 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvz7" event={"ID":"69047018-e836-408d-9ff4-f6d9ee4a37b2","Type":"ContainerDied","Data":"18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32"} Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.144499 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrvz7" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.144532 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvz7" event={"ID":"69047018-e836-408d-9ff4-f6d9ee4a37b2","Type":"ContainerDied","Data":"f45caaea490893f7fa52af61c5e11466b24ca3554bf9aff6751d648751c80151"} Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.144553 4948 scope.go:117] "RemoveContainer" containerID="18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.168434 4948 scope.go:117] "RemoveContainer" containerID="69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.210192 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hrvz7"] Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.211396 4948 scope.go:117] "RemoveContainer" containerID="3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.230008 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hrvz7"] Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.264341 4948 scope.go:117] "RemoveContainer" containerID="18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32" Feb 20 09:16:44 crc kubenswrapper[4948]: E0220 09:16:44.264776 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32\": container with ID starting with 18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32 not found: ID does not exist" containerID="18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.264806 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32"} err="failed to get container status \"18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32\": rpc error: code = NotFound desc = could not find container \"18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32\": container with ID starting with 18ea1ab3b95b4d88bfcc704c5faf23f222ba4425d7abde000ef461eb010ddd32 not found: ID does not exist" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.264831 4948 scope.go:117] "RemoveContainer" containerID="69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758" Feb 20 09:16:44 crc kubenswrapper[4948]: E0220 09:16:44.265319 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758\": container with ID starting with 69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758 not found: ID does not exist" containerID="69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.265420 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758"} err="failed to get container status \"69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758\": rpc error: code = NotFound desc = could not find container \"69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758\": container with ID starting with 69f6a62bc42ec81d9c5bc2e4e325ac62148fca78690de6243134a1c2fa0d1758 not found: ID does not exist" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.265495 4948 scope.go:117] "RemoveContainer" containerID="3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438" Feb 20 09:16:44 crc kubenswrapper[4948]: E0220 09:16:44.265834 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438\": container with ID starting with 3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438 not found: ID does not exist" containerID="3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438" Feb 20 09:16:44 crc kubenswrapper[4948]: I0220 09:16:44.265916 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438"} err="failed to get container status \"3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438\": rpc error: code = NotFound desc = could not find container \"3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438\": container with ID starting with 3a5d3c0f1a659e4e232e276ac70ee5562e34ae0a4e76e4f6f29a8672c1834438 not found: ID does not exist" Feb 20 09:16:45 crc kubenswrapper[4948]: I0220 09:16:45.736851 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69047018-e836-408d-9ff4-f6d9ee4a37b2" path="/var/lib/kubelet/pods/69047018-e836-408d-9ff4-f6d9ee4a37b2/volumes" Feb 20 09:16:52 crc kubenswrapper[4948]: I0220 09:16:52.722833 4948 scope.go:117] "RemoveContainer" containerID="2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5" Feb 20 09:16:52 crc kubenswrapper[4948]: E0220 09:16:52.724000 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:17:04 crc kubenswrapper[4948]: I0220 09:17:04.722680 4948 scope.go:117] "RemoveContainer" containerID="2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5" Feb 20 09:17:04 crc kubenswrapper[4948]: E0220 09:17:04.723616 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:17:18 crc kubenswrapper[4948]: I0220 09:17:18.723552 4948 scope.go:117] "RemoveContainer" containerID="2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5" Feb 20 09:17:18 crc kubenswrapper[4948]: E0220 09:17:18.724414 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" Feb 20 09:17:33 crc kubenswrapper[4948]: I0220 09:17:33.732286 4948 scope.go:117] "RemoveContainer" containerID="2e609219eda64b1acca0e9ec84aa03d7c21986cd69e1f5ce367d12346cb500e5" Feb 20 09:17:33 crc kubenswrapper[4948]: E0220 09:17:33.733346 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h4ww2_openshift-machine-config-operator(84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee)\"" pod="openshift-machine-config-operator/machine-config-daemon-h4ww2" podUID="84f9cd7f-aa8a-457c-9cc7-8eb65fb4b2ee" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515146023270024445 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015146023271017363 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015146012416016505 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015146012417015456 5ustar corecore